Music Similarity Representation Learning Focusing on Individual Instruments with Source Separation and Human Preference
By: Takehiro Imamura , Yuka Hashizume , Wen-Chin Huang and more
Potential Business Impact:
Helps computers find songs that sound alike.
This paper proposes music similarity representation learning (MSRL) based on individual instrument sounds (InMSRL) utilizing music source separation (MSS) and human preference without requiring clean instrument sounds during inference. We propose three methods that effectively improve performance. First, we introduce end-to-end fine-tuning (E2E-FT) for the Cascade approach that sequentially performs MSS and music similarity feature extraction. E2E-FT allows the model to minimize the adverse effects of a separation error on the feature extraction. Second, we propose multi-task learning for the Direct approach that directly extracts disentangled music similarity features using a single music similarity feature extractor. Multi-task learning, which is based on the disentangled music similarity feature extraction and MSS based on reconstruction with disentangled music similarity features, further enhances instrument feature disentanglement. Third, we employ perception-aware fine-tuning (PAFT). PAFT utilizes human preference, allowing the model to perform InMSRL aligned with human perceptual similarity. We conduct experimental evaluations and demonstrate that 1) E2E-FT for Cascade significantly improves InMSRL performance, 2) the multi-task learning for Direct is also helpful to improve disentanglement performance in the feature extraction, 3) PAFT significantly enhances the perceptual InMSRL performance, and 4) Cascade with E2E-FT and PAFT outperforms Direct with the multi-task learning and PAFT.
Similar Papers
Learning Separated Representations for Instrument-based Music Similarity
Sound
Find songs by their instruments, not just the whole song.
Musical Source Separation of Brazilian Percussion
Audio and Speech Processing
Separates samba drums from music using smart computer.
Training a Perceptual Model for Evaluating Auditory Similarity in Music Adversarial Attack
Sound
Makes music AI understand songs like people do.