Joint Estimation of Piano Dynamics and Metrical Structure with a Multi-task Multi-Scale Network
By: Zhanhong He , Hanyu Meng , David Huang and more
Potential Business Impact:
Helps computers understand piano music's loudness.
Estimating piano dynamic from audio recordings is a fundamental challenge in computational music analysis. In this paper, we propose an efficient multi-task network that jointly predicts dynamic levels, change points, beats, and downbeats from a shared latent representation. These four targets form the metrical structure of dynamics in the music score. Inspired by recent vocal dynamic research, we use a multi-scale network as the backbone, which takes Bark-scale specific loudness as the input feature. Compared to log-Mel as input, this reduces model size from 14.7 M to 0.5 M, enabling long sequential input. We use a 60-second audio length in audio segmentation, which doubled the length of beat tracking commonly used. Evaluated on the public MazurkaBL dataset, our model achieves state-of-the-art results across all tasks. This work sets a new benchmark for piano dynamic estimation and delivers a powerful and compact tool, paving the way for large-scale, resource-efficient analysis of musical expression.
Similar Papers
Evaluating High-Resolution Piano Sustain Pedal Depth Estimation with Musically Informed Metrics
Information Retrieval
Helps music programs understand piano pedal use better.
Exploring System Adaptations For Minimum Latency Real-Time Piano Transcription
Audio and Speech Processing
Lets computers hear piano music instantly.
PianoBind: A Multimodal Joint Embedding Model for Pop-piano Music
Sound
Helps computers understand piano music's feelings.