INFNet: A Task-aware Information Flow Network for Large-Scale Recommendation Systems
By: Kaiyuan Li , Dongdong Mao , Yongxiang Tang and more
Potential Business Impact:
Helps online ads show better things you'll like.
Feature interaction has long been a cornerstone of ranking models in large-scale recommender systems due to its proven effectiveness in capturing complex dependencies among features. However, existing feature interaction strategies face two critical challenges in industrial applications: (1) The vast number of categorical and sequential features makes exhaustive interaction computationally prohibitive, often resulting in optimization difficulties. (2) Real-world recommender systems typically involve multiple prediction objectives, yet most current approaches apply feature interaction modules prior to the multi-task learning layers. This late-fusion design overlooks task-specific feature dependencies and inherently limits the capacity of multi-task modeling. To address these limitations, we propose the Information Flow Network (INFNet), a task-aware architecture designed for large-scale recommendation scenarios. INFNet distinguishes features into three token types, categorical tokens, sequence tokens, and task tokens, and introduces a novel dual-flow design comprising heterogeneous and homogeneous alternating information blocks. For heterogeneous information flow, we employ a cross-attention mechanism with proxy that facilitates efficient cross-modal token interaction with balanced computational cost. For homogeneous flow, we design type-specific Proxy Gated Units (PGUs) to enable fine-grained intra-type feature processing. Extensive experiments on multiple offline benchmarks confirm that INFNet achieves state-of-the-art performance. Moreover, INFNet has been successfully deployed in a commercial online advertising system, yielding significant gains of +1.587% in Revenue (REV) and +1.155% in Click-Through Rate (CTR).
Similar Papers
ViFusion: In-Network Tensor Fusion for Scalable Video Feature Indexing
Multimedia
Speeds up finding videos by 8 to 22 times.
Information Filtering Networks: Theoretical Foundations, Generative Methodologies, and Real-World Applications
Machine Learning (CS)
Finds hidden connections in complicated information.
AIF: Asynchronous Inference Framework for Cost-Effective Pre-Ranking
Machine Learning (CS)
Makes online ads show faster and better.