HeartMuLa: A Family of Open Sourced Music Foundation Models
By: Dongchao Yang , Yuxin Xie , Yuguo Yin and more
We present a family of open-source Music Foundation Models designed to advance large-scale music understanding and generation across diverse tasks and modalities. Our framework consists of four major components: (1) HeartCLAP, an audio-text alignment model; (2) HeartTranscriptor, a robust lyric recognition model optimized for real-world music scenarios; and (3) HeartCodec, a low-frame-rate (12.5 Hz) yet high-fidelity music codec tokenizer that captures long-range musical structure while preserving fine-grained acoustic details and enabling efficient autoregressive modeling; (4) HeartMuLa, an LLM-based song generation model capable of synthesizing high-fidelity music under rich, user-controllable conditions (e.g., textual style descriptions, lyrics, and reference audio). In addition, it provides two specialized modes: (i) fine-grained musical attribute control, which allows users to specify the style of different song sections (e.g., intro, verse, chorus) using natural language prompts; and (ii) short, engaging music generation, which is suitable as background music for short videos. Lastly, HeartMuLa improves significantly when scaled to 7B parameters. For the first time, we show that a Suno-level, commercial-grade system can be reproduced using academic-scale data and GPU resources. We expect these foundation models to serve as strong baselines for future research and to facilitate practical applications in multimodal content production.
Similar Papers
Exploring Finetuned Audio-LLM on Heart Murmur Features
Audio and Speech Processing
Helps doctors hear heart problems better.
Muse: Towards Reproducible Long-Form Song Generation with Fine-Grained Style Control
Sound
Lets anyone make songs with music and words.
Muse: Towards Reproducible Long-Form Song Generation with Fine-Grained Style Control
Sound
Makes computers create songs with any style.