Published November 3, 2025 | Version v1
Conference paper Open

A Study on the Data Distribution Gap in Music Emotion Recognition

Description

Music Emotion Recognition (MER) is a task deeply connected to human perception, relying heavily on subjective annotations collected from contributors. Prior studies tend to focus on specific musical styles rather than incorporating a diverse range of genres, such as rock and classical, within a single framework. In this paper, we address the task of recognizing emotion from audio content by investigating five datasets with dimensional emotion annotations - EmoMusic, DEAM, PMEmo, WTC, and WCMED - which span various musical styles. We demonstrate the problem of out-of-distribution generalization in a systematic experiment. By closely looking at multiple data and feature sets, we provide insight into genre-emotion relationships in existing data and examine potential genre dominance and dataset biases in certain feature representations. Based on these experiments, we arrive at a simple yet effective framework that combines embeddings extracted from the Jukebox model with chroma features and demonstrate how, alongside a combination of several diverse training sets, this permits us to train models with substantially improved cross-dataset generalization capabilities.

Files

CMMR2025_O7_4.pdf

Files (3.0 MB)

Name Size Download all
md5:cff863ffe2bb4a04f41072dfb151861e
3.0 MB Preview Download