Context-Gated Cross-Modal Perception with Visual Mamba for PET-CT Lung Tumor Segmentation
By: Elena Mulero Ayllón , Linlin Shen , Pierangelo Veltri and more
Potential Business Impact:
Finds lung tumors better using two types of scans.
Accurate lung tumor segmentation is vital for improving diagnosis and treatment planning, and effectively combining anatomical and functional information from PET and CT remains a major challenge. In this study, we propose vMambaX, a lightweight multimodal framework integrating PET and CT scan images through a Context-Gated Cross-Modal Perception Module (CGM). Built on the Visual Mamba architecture, vMambaX adaptively enhances inter-modality feature interaction, emphasizing informative regions while suppressing noise. Evaluated on the PCLT20K dataset, the model outperforms baseline models while maintaining lower computational complexity. These results highlight the effectiveness of adaptive cross-modal gating for multimodal tumor segmentation and demonstrate the potential of vMambaX as an efficient and scalable framework for advanced lung cancer analysis. The code is available at https://github.com/arco-group/vMambaX.
Similar Papers
Cross-Modal Interactive Perception Network with Mamba for Lung Tumor Segmentation in PET-CT Images
Image and Video Processing
Finds lung tumors better in body scans.
HybridMamba: A Dual-domain Mamba for 3D Medical Image Segmentation
CV and Pattern Recognition
Helps doctors see inside bodies better.
CECT-Mamba: a Hierarchical Contrast-enhanced-aware Model for Pancreatic Tumor Subtyping from Multi-phase CECT
CV and Pattern Recognition
Helps doctors tell tumor types apart better.