Towards Reliable Objective Evaluation Metrics for Generative Singing Voice Separation Models
Authors/Creators
Description
This upload accompanies the WASPAA 2025 paper, 'Towards Reliable Objective Evaluation Metrics for Generative Singing Voice Separation Models'.
It contains the evaluation audio used to compute the evaluation metrics, as well as the loudness-normalised test stimuli used in the DCR test.
Also included is a CSV file containing all the metrics and DMOS scores used to calculate Spearman's rank-based correlation coefficients (SRCCs) for evaluating both discriminative and generative models.
A demonstration Python script outlines how the SRCCs between the DMOS and the evaluation metrics for both types of model are calculated.
The Readme.md file provides step-by-step instructions on how to benchmark metrics not included in the analysis.
Files
gensvs_eval_data.zip
Files
(810.4 MB)
| Name | Size | Download all |
|---|---|---|
|
md5:3790b0f47f56eae10fc9604b3f7a1011
|
810.4 MB | Preview Download |
Additional details
Additional titles
- Subtitle (English)
- Evaluation Data: Audio, Embeddings, Metrics and DMOS
Dates
- Available
-
2025-07-15
Software
- Repository URL
- https://github.com/pablebe/gensvs_eval
- Programming language
- Python
- Development Status
- Active
References
- gensvs_eval