Automatic Music Transcription using Convolutional Neural Networks and Constant-Q transform
By: Yohannis Telila, Tommaso Cucinotta, Davide Bacciu
Potential Business Impact:
Turns piano music into written notes.
Automatic music transcription (AMT) is the problem of analyzing an audio recording of a musical piece and detecting notes that are being played. AMT is a challenging problem, particularly when it comes to polyphonic music. The goal of AMT is to produce a score representation of a music piece, by analyzing a sound signal containing multiple notes played simultaneously. In this work, we design a processing pipeline that can transform classical piano audio files in .wav format into a music score representation. The features from the audio signals are extracted using the constant-Q transform, and the resulting coefficients are used as an input to the convolutional neural network (CNN) model.
Similar Papers
TART: A Comprehensive Tool for Technique-Aware Audio-to-Tab Guitar Transcription
Sound
Turns guitar music into written notes.
Count The Notes: Histogram-Based Supervision for Automatic Music Transcription
Sound
Turns music into notes with less work.
Sound and Music Biases in Deep Music Transcription Models: A Systematic Analysis
Sound
Helps computers understand music better, not just piano.