FlexiReID: Adaptive Mixture of Expert for Multi-Modal Person Re-Identification
By: Zhen Sun , Lei Tan , Yunhang Shen and more
Potential Business Impact:
Find people in pictures, even with different kinds of clues.
Multimodal person re-identification (Re-ID) aims to match pedestrian images across different modalities. However, most existing methods focus on limited cross-modal settings and fail to support arbitrary query-retrieval combinations, hindering practical deployment. We propose FlexiReID, a flexible framework that supports seven retrieval modes across four modalities: rgb, infrared, sketches, and text. FlexiReID introduces an adaptive mixture-of-experts (MoE) mechanism to dynamically integrate diverse modality features and a cross-modal query fusion module to enhance multimodal feature extraction. To facilitate comprehensive evaluation, we construct CIRS-PEDES, a unified dataset extending four popular Re-ID datasets to include all four modalities. Extensive experiments demonstrate that FlexiReID achieves state-of-the-art performance and offers strong generalization in complex scenarios.
Similar Papers
Identity Clue Refinement and Enhancement for Visible-Infrared Person Re-Identification
CV and Pattern Recognition
Helps cameras find people in different light.
ReID5o: Achieving Omni Multi-modal Person Re-identification in a Single Model
CV and Pattern Recognition
Find people using any pictures or words.
MMReID-Bench: Unleashing the Power of MLLMs for Effective and Versatile Person Re-identification
CV and Pattern Recognition
Find people in pictures using many kinds of clues.