Score: 0

Segment Transformer: AI-Generated Music Detection via Music Structural Analysis

Published: September 10, 2025 | arXiv ID: 2509.08283v1

By: Yumin Kim, Seonghyeon Go

Potential Business Impact:

Tells if music was made by AI or people.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Audio and music generation systems have been remarkably developed in the music information retrieval (MIR) research field. The advancement of these technologies raises copyright concerns, as ownership and authorship of AI-generated music (AIGM) remain unclear. Also, it can be difficult to determine whether a piece was generated by AI or composed by humans clearly. To address these challenges, we aim to improve the accuracy of AIGM detection by analyzing the structural patterns of music segments. Specifically, to extract musical features from short audio clips, we integrated various pre-trained models, including self-supervised learning (SSL) models or an audio effect encoder, each within our suggested transformer-based framework. Furthermore, for long audio, we developed a segment transformer that divides music into segments and learns inter-segment relationships. We used the FakeMusicCaps and SONICS datasets, achieving high accuracy in both the short-audio and full-audio detection experiments. These findings suggest that integrating segment-level musical features into long-range temporal analysis can effectively enhance both the performance and robustness of AIGM detection systems.

Page Count
6 pages

Category
Computer Science:
Sound