Score: 0

Context-Gated Cross-Modal Perception with Visual Mamba for PET-CT Lung Tumor Segmentation

Published: October 31, 2025 | arXiv ID: 2510.27508v1

By: Elena Mulero Ayllón , Linlin Shen , Pierangelo Veltri and more

Potential Business Impact:

Finds lung tumors better using two types of scans.

Business Areas:
Image Recognition Data and Analytics, Software

Accurate lung tumor segmentation is vital for improving diagnosis and treatment planning, and effectively combining anatomical and functional information from PET and CT remains a major challenge. In this study, we propose vMambaX, a lightweight multimodal framework integrating PET and CT scan images through a Context-Gated Cross-Modal Perception Module (CGM). Built on the Visual Mamba architecture, vMambaX adaptively enhances inter-modality feature interaction, emphasizing informative regions while suppressing noise. Evaluated on the PCLT20K dataset, the model outperforms baseline models while maintaining lower computational complexity. These results highlight the effectiveness of adaptive cross-modal gating for multimodal tumor segmentation and demonstrate the potential of vMambaX as an efficient and scalable framework for advanced lung cancer analysis. The code is available at https://github.com/arco-group/vMambaX.

Country of Origin
🇮🇹 Italy

Page Count
4 pages

Category
Computer Science:
CV and Pattern Recognition