LeMoF: Level-guided Multimodal Fusion for Heterogeneous Clinical Data
By: Jongseok Kim , Seongae Kang , Jonghwan Shin and more
Multimodal clinical prediction is widely used to integrate heterogeneous data such as Electronic Health Records (EHR) and biosignals. However, existing methods tend to rely on static modality integration schemes and simple fusion strategies. As a result, they fail to fully exploit modality-specific representations. In this paper, we propose Level-guided Modal Fusion (LeMoF), a novel framework that selectively integrates level-guided representations within each modality. Each level refers to a representation extracted from a different layer of the encoder. LeMoF explicitly separates and learns global modality-level predictions from level-specific discriminative representations. This design enables LeMoF to achieve a balanced performance between prediction stability and discriminative capability even in heterogeneous clinical environments. Experiments on length of stay prediction using Intensive Care Unit (ICU) data demonstrate that LeMoF consistently outperforms existing state-of-the-art multimodal fusion techniques across various encoder configurations. We also confirmed that level-wise integration is a key factor in achieving robust predictive performance across various clinical conditions.
Similar Papers
MoE-Health: A Mixture of Experts Framework for Robust Multimodal Healthcare Prediction
Machine Learning (CS)
Helps doctors predict sickness with mixed patient data.
A Unified Framework for Emotion Recognition and Sentiment Analysis via Expert-Guided Multimodal Fusion with Large Language Models
Computation and Language
**Computers understand feelings from talking, seeing, and writing.**
Meta Fusion: A Unified Framework For Multimodality Fusion with Mutual Learning
Machine Learning (CS)
Combines different data to make better predictions.