Worrall, Kyle orcid.org/0000-0001-8600-8430, Yin, Zongyu orcid.org/0000-0001-8709-8829 and Collins, Tom orcid.org/0000-0001-7880-5093 (2024) Comparative evaluation in the wild: Systems for the expressive rendering of music. Transactions on Artificial Intelligence. (In Press)
Abstract
There have been many attempts to model the ability of human musicians to take a score and perform or render it expressively, by adding tempo, timing, loudness and articulation changes to non-expressive music data. While expressive rendering models exist in academic research, most of these are not open source or accessible, meaning they are difficult to evaluate empirically and have not been widely adopted in professional music software. Systematic comparative evaluation of such algorithms stopped after the last Performance Rendering Contest (RENCON) in 2013, making it difficult to compare newer models to existing work in a fair and valid way. In this paper, we introduce the first transformer-based model for expressive rendering, Cue-Free Express + Pedal (CFE+P), which predicts expressive attributes such as note-wise dynamics and micro�timing adjustments, and beat-wise tempo and sustain pedal use based only on the start and end times and pitches of notes (e.g., in�expressive MIDI input). We perform two comparative evaluations on our model against a non-machine learning baseline taken from professional music software and two open-source algorithms – a feedforward neural network (FFNN) and hierarchical recurrent neural network (HRNN). The results of two listening studies indicate that our model renders passages that outperform what can be done in professional music software such as Logic Pro and Ableton Live.
Metadata
Item Type: | Article |
---|---|
Authors/Creators: |
|
Copyright, Publisher and Additional Information: | © IEEE 2024. This is an author-produced version of the published paper. Uploaded with permission of the publisher/copyright holder. Further copying may not be permitted; contact the publisher for details |
Keywords: | Artificial Intelligence in art and music, Computer Generated Music, Music Information Retrieval, Neural Networks, Deep Learning. |
Dates: |
|
Institution: | The University of York |
Funding Information: | Funder Grant number ESPRC CDT for Intelligent Games and Game Intelligence [EP/S022325/1] |
Depositing User: | Mr Kyle Worrall |
Date Deposited: | 25 Sep 2024 14:28 |
Last Modified: | 25 Sep 2024 14:28 |
Published Version: | https://ieeexplore.ieee.org/abstract/document/1054... |
Status: | In Press |
Publisher: | IEEE |
Related URLs: | |
Open Archives Initiative ID (OAI ID): | oai:eprints.whiterose.ac.uk:217634 |