Exploring Procedural Data Generation for Automatic Acoustic Guitar Fingerpicking Transcription
By: Sebastian Murgul, Michael Heizmann
Potential Business Impact:
Makes computers learn guitar music from fake sounds.
Automatic transcription of acoustic guitar fingerpicking performances remains a challenging task due to the scarcity of labeled training data and legal constraints connected with musical recordings. This work investigates a procedural data generation pipeline as an alternative to real audio recordings for training transcription models. Our approach synthesizes training data through four stages: knowledge-based fingerpicking tablature composition, MIDI performance rendering, physical modeling using an extended Karplus-Strong algorithm, and audio augmentation including reverb and distortion. We train and evaluate a CRNN-based note-tracking model on both real and synthetic datasets, demonstrating that procedural data can be used to achieve reasonable note-tracking results. Finetuning with a small amount of real data further enhances transcription accuracy, improving over models trained exclusively on real recordings. These results highlight the potential of procedurally generated audio for data-scarce music information retrieval tasks.
Similar Papers
A Machine Learning Approach for MIDI to Guitar Tablature Conversion
Sound
Turns music into guitar finger instructions.
Joint Transcription of Acoustic Guitar Strumming Directions and Chords
Sound
Helps computers understand guitar strumming and chords.
TART: A Comprehensive Tool for Technique-Aware Audio-to-Tab Guitar Transcription
Sound
Turns guitar music into written notes.