Published June 30, 2021 | Version v1
Conference paper Open

Tracing Back Music Emotion Predictions to Sound Sources and Intuitive Perceptual Qualities

  • 1. Institute of Computational Perception, Johannes Kepler University, Linz, Austria
  • 2. Institute of Computational Perception, Johannes Kepler University, Linz, Austria - LIT AI Lab, Linz Institute of Technology, Linz, Austria

Description

Music emotion recognition is an important task in MIR (Music Information Retrieval) research. Owing to factors like the subjective nature of the task and the variation of emotional cues between musical genres, there are still significant challenges in developing reliable and generalizable models. One important step towards better models would be to understand what a model is actually learning from the data and how the prediction for a particular input is made. In previous work, we have shown how to derive explanations of model predictions in terms of spectrogram image segments that connect to the high-level emotion prediction via a layer of easily interpretable perceptual features. However, that scheme lacks intuitive musical comprehensibility at the spectrogram level. In the present work, we bridge this gap by merging audioLIME – a source-separation based explainer – with mid-level perceptual features, thus forming an intuitive connection chain between the input audio and the output emotion predictions. We demonstrate the usefulness of this method by applying it to debug a biased emotion prediction model.

Files

SMC_2021_paper_87.pdf

Files (813.3 kB)

Name Size Download all
md5:13e6bfa567b42e9fc4d20b8f7b35bcff
813.3 kB Preview Download