SMART: Tuning a symbolic music generation system with an audio domain aesthetic reward
Description
Recent work has proposed training machine learning models to predict aesthetic ratings for music audio. Our work explores whether such models can be used to finetune a symbolic music generation system with reinforcement learning, and what effect this has on the system outputs. To test this, we use group relative policy optimization to finetune a piano MIDI model with Meta Audiobox Aesthetics ratings of audio-rendered outputs as the reward. We find that this optimization affects multiple low-level features of the generated outputs, and improves the average subjective ratings in a preliminary listening study with 14 participants. We also find that over-optimization dramatically reduces diversity of model outputs. Code and listening examples can be found here: https://github.com/erl-j/SMART.
Files
_SMART (7).pdf
Files
(962.5 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:57c7a2750f5f66ea0dca739eb654c44f
|
962.5 kB | Preview Download |