Reverse-Engineering the Retrieval Process in GenIR Models
By: Anja Reusch, Yonatan Belinkov
Potential Business Impact:
Shows how computers pick the best search results.
Generative Information Retrieval (GenIR) is a novel paradigm in which a transformer encoder-decoder model predicts document rankings based on a query in an end-to-end fashion. These GenIR models have received significant attention due to their simple retrieval architecture while maintaining high retrieval effectiveness. However, in contrast to established retrieval architectures like cross-encoders or bi-encoders, their internal computations remain largely unknown. Therefore, this work studies the internal retrieval process of GenIR models by applying methods based on mechanistic interpretability, such as patching and vocabulary projections. By replacing the GenIR encoder with one trained on fewer documents, we demonstrate that the decoder is the primary component responsible for successful retrieval. Our patching experiments reveal that not all components in the decoder are crucial for the retrieval process. More specifically, we find that a pass through the decoder can be divided into three stages: (I) the priming stage, which contributes important information for activating subsequent components in later layers; (II) the bridging stage, where cross-attention is primarily active to transfer query information from the encoder to the decoder; and (III) the interaction stage, where predominantly MLPs are active to predict the document identifier. Our findings indicate that interaction between query and document information occurs only in the last stage. We hope our results promote a better understanding of GenIR models and foster future research to overcome the current challenges associated with these models.
Similar Papers
GenIR: Generative Visual Feedback for Mental Image Retrieval
CV and Pattern Recognition
Helps find pictures from your thoughts.
Lightweight and Direct Document Relevance Optimization for Generative Information Retrieval
Information Retrieval
Finds better search results without complex math.
ImpRAG: Retrieval-Augmented Generation with Implicit Queries
Computation and Language
Lets computers find answers without being asked.