Score: 1

Topologic Attention Networks: Attending to Direct and Indirect Neighbors through Gaussian Belief Propagation

Published: November 21, 2025 | arXiv ID: 2511.16871v1

By: Marshall Rosenhoover, Huaming Zhang

Potential Business Impact:

Lets computers understand complex connections faster.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Graph Neural Networks rely on local message passing, which limits their ability to model long-range dependencies in graphs. Existing approaches extend this range through continuous-time dynamics or dense self-attention, but both suffer from high computational cost and limited scalability. We propose Topologic Attention Networks, a new framework that applies topologic attention, a probabilistic mechanism that learns how information should flow through both direct and indirect connections in a graph. Unlike conventional attention that depends on explicit pairwise interactions, topologic attention emerges from the learned information propagation of the graph, enabling unified reasoning over local and global relationships. This method achieves provides state-of-the-art performance across all measured baseline models. Our implementation is available at https://github.com/Marshall-Rosenhoover/Topologic-Attention-Networks.

Country of Origin
🇺🇸 United States

Page Count
15 pages

Category
Computer Science:
Machine Learning (CS)