Published June 30, 2021 | Version v1
Conference paper Open

How Does the Spotify API Compare to the Music Emotion Recognition State-of-the-Art?

  • 1. CISUC, DEI, University of Coimbra, Portugal - Ci2,Smart Cities Research Center, Instituto Politécnico de Tomar, Tomar, Portugal
  • 2. CISUC, DEI, University of Coimbra, Portugal
  • 3. CISUC, DEI, University of Coimbra, Portugal - Miguel Torga Higher Institute, Coimbra, Portugal

Description

Features are arguably the key factor to any machine learning problem. Over the decades, myriads of audio features and recently feature-learning approaches have been tested in Music Emotion Recognition (MER) with scarce improvements. Here, we shed some light on the suitability of the audio features provided by the Spotify API, the leading music streaming service, when applied to MER. To this end, 12 Spotify API features were obtained for 704 of our 900-song dataset, annotated in terms of Russell’s quadrants. These are compared to emotionally-relevant features obtained previously, using feature ranking and emotion classification experiments. We verified that energy, valence and acousticness features from Spotify are highly relevant to MER. However, the 12-feature set is unable to meet the performance of the features available in the stateof- the-art (58.5% vs. 74.7% F1-measure). Combining Spotify and state-of-the-art sets leads to small improvements with fewer features (top5: +2.3%, top10: +1.1%), while not improving the highest results (100 features). From this we conclude that Spotify provides some higherlevel emotionally-relevant features. Such extractors are desirable, since they are closer to human concepts and allow for interpretable rules to be extracted (harder with hundreds of abstract features). Still, additional emotionallyrelevant features are needed to improve MER.

Files

SMC_2021_paper_84.pdf

Files (994.6 kB)

Name Size Download all
md5:0bcda8d2662e7ec89f70829a2271e485
994.6 kB Preview Download