Hyperbolic Residual Quantization: Discrete Representations for Data with Latent Hierarchies
By: Piotr Piękos, Subhradeep Kayal, Alexandros Karatzoglou
Potential Business Impact:
Maps complex data better for computers.
Hierarchical data arise in countless domains, from biological taxonomies and organizational charts to legal codes and knowledge graphs. Residual Quantization (RQ) is widely used to generate discrete, multitoken representations for such data by iteratively quantizing residuals in a multilevel codebook. However, its reliance on Euclidean geometry can introduce fundamental mismatches that hinder modeling of hierarchical branching, necessary for faithful representation of hierarchical data. In this work, we propose Hyperbolic Residual Quantization (HRQ), which embeds data natively in a hyperbolic manifold and performs residual quantization using hyperbolic operations and distance metrics. By adapting the embedding network, residual computation, and distance metric to hyperbolic geometry, HRQ imparts an inductive bias that aligns naturally with hierarchical branching. We claim that HRQ in comparison to RQ can generate more useful for downstream tasks discrete hierarchical representations for data with latent hierarchies. We evaluate HRQ on two tasks: supervised hierarchy modeling using WordNet hypernym trees, where the model is supervised to learn the latent hierarchy - and hierarchy discovery, where, while latent hierarchy exists in the data, the model is not directly trained or evaluated on a task related to the hierarchy. Across both scenarios, HRQ hierarchical tokens yield better performance on downstream tasks compared to Euclidean RQ with gains of up to $20\%$ for the hierarchy modeling task. Our results demonstrate that integrating hyperbolic geometry into discrete representation learning substantially enhances the ability to capture latent hierarchies.
Similar Papers
A Hierarchical Quantized Tokenization Framework for Task-Adaptive Graph Representation Learning
Information Retrieval
Helps computers understand and learn from complex data.
Distribution-Aware Hadamard Quantization for Hardware-Efficient Implicit Neural Representations
CV and Pattern Recognition
Makes AI models smaller and faster.
Graph is a Natural Regularization: Revisiting Vector Quantization for Graph Representation Learning
Machine Learning (CS)
Helps computers understand complex data better.