P4-07: ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performance
Zhang, Huan*, Tang, Jingjing, Rafee, Syed RM, Dixon, Simon, Fazekas, George, Wiggins, Geraint A.
Subjects (starting with primary): Evaluation, datasets, and reproducibility -> novel datasets and use cases ; Musical features and properties -> expression and performative aspects of music ; MIR fundamentals and methodology -> symbolic music processing ; MIR fundamentals and methodology -> metadata, tags, linked data, and semantic web ; Applications -> performance, and production ; Domain knowledge -> machine learning/artificial intelligence for music
Presented In-person, in Bengaluru: 4-minute short-format presentation
Computational models of expressive piano performance rely on attributes like tempo, timing, dynamics and pedalling. Despite some promising models for performance assessment and performance rendering, results are limited by the scale, breadth and uniformity of existing datasets. In this paper, we present ATEPP, a dataset that contains 1000 hours of performances of standard piano repertoire by 49 world-renowned pianists, organized and aligned by compositions and movements for comparative studies. Scores in MusicXML format are also available for around half of the tracks. We first evaluate and verify the use of transcribed MIDI for representing expressive performance with a listening evaluation that involves recent transcription models. Then, the process of sourcing and curating the dataset is outlined, including composition entity resolution and a pipeline for audio matching and solo filtering. Finally, we conduct baseline experiments for performer identification and performance rendering on our datasets, demonstrating its potential in generalizing expressive features of individual performing style.