Video/Audio Open Access


Turpault, Nicolas; Serizel, Romain

Citation Style Language JSON Export

  "publisher": "Zenodo", 
  "DOI": "10.5281/zenodo.4569096", 
  "title": "DESED_synthetic", 
  "issued": {
    "date-parts": [
  "abstract": "<p>Link to the associated github repository: <a href=\"\"></a></p>\n\n<p>Link to the papers: <a href=\"\"><em></em></a>,&nbsp; <a href=\"\"></a></p>\n\n<p>Domestic Environment Sound Event Detection (DESED).</p>\n\n<p><strong>Description</strong><br>\nThis dataset is the synthetic part of the DESED dataset. It allows creating mixtures of isolated sounds and backgrounds.</p>\n\n<p>There is the material to:</p>\n\n<ul>\n\t<li>Reproduce the DCASE 2019 task 4 synthetic dataset</li>\n\t<li>Reproduce the DCASE 2020 task 4 synthetic train dataset</li>\n\t<li>Creating new mixtures from isolated foreground sounds and background sounds.</li>\n</ul>\n\n<p><strong>Files:</strong></p>\n\n<p><strong>If you want to generate new audio mixtures yourself from the original files.</strong></p>\n\n<ol>\n\t<li><strong>DESED_synth_soundbank.tar.gz</strong> : Raw data used to generate mixtures.</li>\n\t<li><strong>DESED_synth_dcase2019jams.tar.gz</strong>: JAMS files, metadata describing how to recreate the&nbsp; dcase2019 synthetic dataset<strong> </strong></li>\n\t<li><strong>DESED_synth_dcase20_train_val_jams.tar: </strong>JAMS files, metadata describing how to recreate the dcase2020 synthetic train and valid dataset.</li>\n\t<li><strong>DESED_synth_dcase20_eval_jams.tar: </strong>JAMS files, metadata describing how to recreate the dcase2020 synthetic eval dataset (only the basic one, variants of it have been made but not presented here).</li>\n</ol>\n\n<p><strong>If you simply want the evaluation synthetic dataset used in DCASE 2019 task 4.</strong></p>\n\n<ol>\n\t<li><strong>DESED_synth_eval_dcase2019.tar.gz</strong><strong> </strong>:<strong> </strong>Evaluation audio and metadata files used in dcase 2019 task 4.</li>\n</ol>\n\n<p>&nbsp;</p>\n\n<p>The mixtures are generated using Scaper ( [1].</p>\n\n<p>* Background files are extracted from SINS [2], MUSAN [3] or Youtube and have been selected because they contain a very low amount of our sound event classes.<br>\n* Foreground files are extracted from Freesound [4][5] and manually verified to check the quality and segmented to remove silences.</p>\n\n<p><strong>References</strong><br>\n[1] J. Salamon, D. MacConnell, M. Cartwright, P. Li, and J. P. Bello. Scaper: A library for soundscape synthesis and augmentation<br>\nIn IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, Oct. 2017.</p>\n\n<p>[2] Gert Dekkers, Steven Lauwereins, Bart Thoen, Mulu Weldegebreal Adhana, Henk Brouckxon, Toon van Waterschoot, Bart Vanrumste, Marian Verhelst, and Peter Karsmakers.<br>\nThe SINS database for detection of daily activities in a home environment using an acoustic sensor network.<br>\nIn Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 Workshop (DCASE2017), 32&ndash;36. November 2017.</p>\n\n<p>[3] David Snyder and Guoguo Chen and Daniel Povey.<br>\nMUSAN: A Music, Speech, and Noise Corpus.<br>\narXiv, 1510.08484, 2015.</p>\n\n<p>[4] F. Font, G. Roma &amp; X. Serra. Freesound technical demo. In Proceedings of the 21st ACM international conference on Multimedia. ACM, 2013.<br>\n&nbsp;<br>\n[5] E. Fonseca, J. Pons, X. Favory, F. Font, D. Bogdanov, A. Ferraro, S. Oramas, A. Porter &amp; X. Serra. Freesound Datasets: A Platform for the Creation of Open Audio Datasets.<br>\nIn Proceedings of the 18th International Society for Music Information Retrieval Conference, Suzhou, China, 2017.</p>\n\n<p>&nbsp;</p>", 
  "author": [
      "family": "Turpault, Nicolas"
      "family": "Serizel, Romain"
  "version": "v3.0", 
  "type": "motion_picture", 
  "id": "4569096"
All versions This version
Views 2,609388
Downloads 9,590900
Data volume 32.6 TB6.2 TB
Unique views 1,911317
Unique downloads 4,387565


Cite as