Learning the Structure of Connection Graphs
By: Leonardo Di Nino , Gabriele D'Acunto , Sergio Barbarossa and more
Potential Business Impact:
Finds hidden patterns in connected data.
Connection graphs (CGs) extend traditional graph models by coupling network topology with orthogonal transformations, enabling the representation of global geometric consistency. They play a key role in applications such as synchronization, Riemannian signal processing, and neural sheaf diffusion. In this work, we address the inverse problem of learning CGs directly from observed signals. We propose a principled framework based on maximum pseudo-likelihood under a consistency assumption, which enforces spectral properties linking the connection Laplacian to the underlying combinatorial Laplacian. Based on this formulation, we introduce the Structured Connection Graph Learning (SCGL) algorithm, a block-optimization procedure over Riemannian manifolds that jointly infers network topology, edge weights, and geometric structure. Our experiments show that SCGL consistently outperforms existing baselines in both topological recovery and geometric fidelity, while remaining computationally efficient.
Similar Papers
Cross-View Topology-Aware Graph Representation Learning
Machine Learning (CS)
Helps computers understand complex data patterns better.
On the Interplay between Graph Structure and Learning Algorithms in Graph Neural Networks
Machine Learning (CS)
Helps computers learn better from connected data.
A Generative Graph Contrastive Learning Model with Global Signal
Machine Learning (CS)
Makes computer learning from networks smarter.