Score: 1

Label-free Motion-Conditioned Diffusion Model for Cardiac Ultrasound Synthesis

Published: December 10, 2025 | arXiv ID: 2512.09418v1

By: Zhe Li , Hadrien Reynaud , Johanna P Müller and more

Potential Business Impact:

Makes heart videos without needing doctors' notes.

Business Areas:
Motion Capture Media and Entertainment, Video

Ultrasound echocardiography is essential for the non-invasive, real-time assessment of cardiac function, but the scarcity of labelled data, driven by privacy restrictions and the complexity of expert annotation, remains a major obstacle for deep learning methods. We propose the Motion Conditioned Diffusion Model (MCDM), a label-free latent diffusion framework that synthesises realistic echocardiography videos conditioned on self-supervised motion features. To extract these features, we design the Motion and Appearance Feature Extractor (MAFE), which disentangles motion and appearance representations from videos. Feature learning is further enhanced by two auxiliary objectives: a re-identification loss guided by pseudo appearance features and an optical flow loss guided by pseudo flow fields. Evaluated on the EchoNet-Dynamic dataset, MCDM achieves competitive video generation performance, producing temporally coherent and clinically realistic sequences without reliance on manual labels. These results demonstrate the potential of self-supervised conditioning for scalable echocardiography synthesis. Our code is available at https://github.com/ZheLi2020/LabelfreeMCDM.

Country of Origin
🇩🇪 Germany

Repos / Data Links

Page Count
10 pages

Category
Computer Science:
CV and Pattern Recognition