Expressive and Scalable Quantum Fusion for Multimodal Learning
By: Tuyen Nguyen , Trong Nghia Hoang , Phi Le Nguyen and more
Potential Business Impact:
Makes computers learn from many things better.
The aim of this paper is to introduce a quantum fusion mechanism for multimodal learning and to establish its theoretical and empirical potential. The proposed method, called the Quantum Fusion Layer (QFL), replaces classical fusion schemes with a hybrid quantum-classical procedure that uses parameterized quantum circuits to learn entangled feature interactions without requiring exponential parameter growth. Supported by quantum signal processing principles, the quantum component efficiently represents high-order polynomial interactions across modalities with linear parameter scaling, and we provide a separation example between QFL and low-rank tensor-based methods that highlights potential quantum query advantages. In simulation, QFL consistently outperforms strong classical baselines on small but diverse multimodal tasks, with particularly marked improvements in high-modality regimes. These results suggest that QFL offers a fundamentally new and scalable approach to multimodal fusion that merits deeper exploration on larger systems.
Similar Papers
Communication Efficient Adaptive Model-Driven Quantum Federated Learning
Machine Learning (CS)
Makes AI learn faster with less data.
When Federated Learning Meets Quantum Computing: Survey and Research Opportunities
Distributed, Parallel, and Cluster Computing
Makes learning faster and safer for computers.
Quantum Federated Learning: Architectural Elements and Future Directions
Quantum Physics
Makes computers learn faster and safer together.