Segment Transformer: AI-Generated Music Detection via Music Structural Analysis
By: Yumin Kim, Seonghyeon Go
Potential Business Impact:
Tells if music was made by AI or people.
Audio and music generation systems have been remarkably developed in the music information retrieval (MIR) research field. The advancement of these technologies raises copyright concerns, as ownership and authorship of AI-generated music (AIGM) remain unclear. Also, it can be difficult to determine whether a piece was generated by AI or composed by humans clearly. To address these challenges, we aim to improve the accuracy of AIGM detection by analyzing the structural patterns of music segments. Specifically, to extract musical features from short audio clips, we integrated various pre-trained models, including self-supervised learning (SSL) models or an audio effect encoder, each within our suggested transformer-based framework. Furthermore, for long audio, we developed a segment transformer that divides music into segments and learns inter-segment relationships. We used the FakeMusicCaps and SONICS datasets, achieving high accuracy in both the short-audio and full-audio detection experiments. These findings suggest that integrating segment-level musical features into long-range temporal analysis can effectively enhance both the performance and robustness of AIGM detection systems.
Similar Papers
Real-world Music Plagiarism Detection With Music Segment Transcription System
Artificial Intelligence
Finds copied music, even if it sounds different.
Training chord recognition models on artificially generated audio
Sound
Makes music AI learn chords from fake songs.
AI-Generated Music Detection and its Challenges
Sound
Finds fake music made by computers.