{
  "access": {
    "embargo": {
      "active": false,
      "reason": null
    },
    "files": "public",
    "record": "public",
    "status": "open"
  },
  "created": "2023-03-08T19:20:47.778307+00:00",
  "custom_fields": {},
  "deletion_status": {
    "is_deleted": false,
    "status": "P"
  },
  "files": {
    "count": 6,
    "enabled": true,
    "entries": {
      "LICENSE": {
        "access": {
          "hidden": false
        },
        "checksum": "md5:1c11108eda7c915172b10c48276cc189",
        "ext": "bin",
        "id": "b8436b11-fde0-475d-a12f-0b200d0c16d7",
        "key": "LICENSE",
        "links": {
          "content": "https://zenodo.org/api/records/7709052/files/LICENSE/content",
          "self": "https://zenodo.org/api/records/7709052/files/LICENSE"
        },
        "metadata": null,
        "mimetype": "application/octet-stream",
        "size": 1194,
        "storage_class": "L"
      },
      "README.md": {
        "access": {
          "hidden": false
        },
        "checksum": "md5:ed8422ff6e3bb036361383c2ab7bbdcc",
        "ext": "md",
        "id": "3d97d86d-2165-4f57-83ea-93b0692fd4e6",
        "key": "README.md",
        "links": {
          "content": "https://zenodo.org/api/records/7709052/files/README.md/content",
          "self": "https://zenodo.org/api/records/7709052/files/README.md"
        },
        "metadata": null,
        "mimetype": "application/octet-stream",
        "size": 27234,
        "storage_class": "L"
      },
      "foa_dev.zip": {
        "access": {
          "hidden": false
        },
        "checksum": "md5:316b834ee6393c22862c314cc3c7ebb0",
        "ext": "zip",
        "id": "1397585c-d8d2-4030-9d70-8fdca2248259",
        "key": "foa_dev.zip",
        "links": {
          "content": "https://zenodo.org/api/records/7709052/files/foa_dev.zip/content",
          "self": "https://zenodo.org/api/records/7709052/files/foa_dev.zip"
        },
        "metadata": null,
        "mimetype": "application/zip",
        "size": 3406085425,
        "storage_class": "L"
      },
      "metadata_dev.zip": {
        "access": {
          "hidden": false
        },
        "checksum": "md5:e73af95a6d5f3f7e009ac6a70804f44a",
        "ext": "zip",
        "id": "05b55017-9bac-4341-a641-7bcf9610b5b9",
        "key": "metadata_dev.zip",
        "links": {
          "content": "https://zenodo.org/api/records/7709052/files/metadata_dev.zip/content",
          "self": "https://zenodo.org/api/records/7709052/files/metadata_dev.zip"
        },
        "metadata": null,
        "mimetype": "application/zip",
        "size": 1165845,
        "storage_class": "L"
      },
      "mic_dev.zip": {
        "access": {
          "hidden": false
        },
        "checksum": "md5:06967bcc8def1580c2425fabc311dbd2",
        "ext": "zip",
        "id": "ac7eb336-3263-4e7d-a6ed-bc8642681ec7",
        "key": "mic_dev.zip",
        "links": {
          "content": "https://zenodo.org/api/records/7709052/files/mic_dev.zip/content",
          "self": "https://zenodo.org/api/records/7709052/files/mic_dev.zip"
        },
        "metadata": null,
        "mimetype": "application/zip",
        "size": 3370049004,
        "storage_class": "L"
      },
      "video_dev.zip": {
        "access": {
          "hidden": false
        },
        "checksum": "md5:b92295e162600adb545ed99bdc0a9c76",
        "ext": "zip",
        "id": "98cc2efe-654d-4340-9fb9-7f2f7ab25a34",
        "key": "video_dev.zip",
        "links": {
          "content": "https://zenodo.org/api/records/7709052/files/video_dev.zip/content",
          "self": "https://zenodo.org/api/records/7709052/files/video_dev.zip"
        },
        "metadata": null,
        "mimetype": "application/zip",
        "size": 4102871214,
        "storage_class": "L"
      }
    },
    "order": [],
    "total_bytes": 10880199916
  },
  "id": "7709052",
  "is_draft": false,
  "is_published": true,
  "links": {
    "access": "https://zenodo.org/api/records/7709052/access",
    "access_grants": "https://zenodo.org/api/records/7709052/access/grants",
    "access_links": "https://zenodo.org/api/records/7709052/access/links",
    "access_request": "https://zenodo.org/api/records/7709052/access/request",
    "access_users": "https://zenodo.org/api/records/7709052/access/users",
    "archive": "https://zenodo.org/api/records/7709052/files-archive",
    "archive_media": "https://zenodo.org/api/records/7709052/media-files-archive",
    "communities": "https://zenodo.org/api/records/7709052/communities",
    "communities-suggestions": "https://zenodo.org/api/records/7709052/communities-suggestions",
    "doi": "https://doi.org/10.5281/zenodo.7709052",
    "draft": "https://zenodo.org/api/records/7709052/draft",
    "file_modification": "https://zenodo.org/api/records/7709052/file-modification",
    "files": "https://zenodo.org/api/records/7709052/files",
    "latest": "https://zenodo.org/api/records/7709052/versions/latest",
    "latest_html": "https://zenodo.org/records/7709052/latest",
    "media_files": "https://zenodo.org/api/records/7709052/media-files",
    "parent": "https://zenodo.org/api/records/7709051",
    "parent_doi": "https://doi.org/10.5281/zenodo.7709051",
    "parent_doi_html": "https://zenodo.org/doi/10.5281/zenodo.7709051",
    "parent_html": "https://zenodo.org/records/7709051",
    "preview_html": "https://zenodo.org/records/7709052?preview=1",
    "quota_increase": "https://zenodo.org/api/records/7709052/quota-increase",
    "request_deletion": "https://zenodo.org/api/records/7709052/request-deletion",
    "requests": "https://zenodo.org/api/records/7709052/requests",
    "reserve_doi": "https://zenodo.org/api/records/7709052/draft/pids/doi",
    "self": "https://zenodo.org/api/records/7709052",
    "self_doi": "https://doi.org/10.5281/zenodo.7709052",
    "self_doi_html": "https://zenodo.org/doi/10.5281/zenodo.7709052",
    "self_html": "https://zenodo.org/records/7709052",
    "self_iiif_manifest": "https://zenodo.org/api/iiif/record:7709052/manifest",
    "self_iiif_sequence": "https://zenodo.org/api/iiif/record:7709052/sequence/default",
    "versions": "https://zenodo.org/api/records/7709052/versions"
  },
  "media_files": {
    "count": 0,
    "enabled": false,
    "entries": {},
    "order": [],
    "total_bytes": 0
  },
  "metadata": {
    "creators": [
      {
        "affiliations": [
          {
            "name": "Tampere University"
          }
        ],
        "person_or_org": {
          "family_name": "Politis",
          "given_name": "Archontis",
          "identifiers": [
            {
              "identifier": "0000-0002-0595-2356",
              "scheme": "orcid"
            }
          ],
          "name": "Politis, Archontis",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "SONY"
          }
        ],
        "person_or_org": {
          "family_name": "Shimada",
          "given_name": "Kazuki",
          "identifiers": [
            {
              "identifier": "0000-0001-5389-2346",
              "scheme": "orcid"
            }
          ],
          "name": "Shimada, Kazuki",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "Tampere University"
          }
        ],
        "person_or_org": {
          "family_name": "Sudarsanam",
          "given_name": "Parthasaarathy",
          "name": "Sudarsanam, Parthasaarathy",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "Tampere University"
          }
        ],
        "person_or_org": {
          "family_name": "Hakala",
          "given_name": "Aapo",
          "name": "Hakala, Aapo",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "SONY"
          }
        ],
        "person_or_org": {
          "family_name": "Takahashi",
          "given_name": "Shusuke",
          "name": "Takahashi, Shusuke",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "Tampere University"
          }
        ],
        "person_or_org": {
          "family_name": "Krause",
          "given_name": "Daniel Alexander",
          "name": "Krause, Daniel Alexander",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "SONY"
          }
        ],
        "person_or_org": {
          "family_name": "Takahashi",
          "given_name": "Naoya",
          "name": "Takahashi, Naoya",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "Tampere University"
          }
        ],
        "person_or_org": {
          "family_name": "Adavanne",
          "given_name": "Sharath",
          "identifiers": [
            {
              "identifier": "0000-0002-5001-6911",
              "scheme": "orcid"
            }
          ],
          "name": "Adavanne, Sharath",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "SONY"
          }
        ],
        "person_or_org": {
          "family_name": "Koyama",
          "given_name": "Yuichiro",
          "name": "Koyama, Yuichiro",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "SONY"
          }
        ],
        "person_or_org": {
          "family_name": "Uchida",
          "given_name": "Kengo",
          "name": "Uchida, Kengo",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "SONY"
          }
        ],
        "person_or_org": {
          "family_name": "Mitsufuji",
          "given_name": "Yuki",
          "identifiers": [
            {
              "identifier": "0000-0002-6806-6140",
              "scheme": "orcid"
            }
          ],
          "name": "Mitsufuji, Yuki",
          "type": "personal"
        }
      },
      {
        "affiliations": [
          {
            "name": "Tampere University"
          }
        ],
        "person_or_org": {
          "family_name": "Virtanen",
          "given_name": "Tuomas",
          "identifiers": [
            {
              "identifier": "0000-0002-4604-9729",
              "scheme": "orcid"
            }
          ],
          "name": "Virtanen, Tuomas",
          "type": "personal"
        }
      }
    ],
    "description": "<p><strong>DESCRIPTION:</strong></p>\n\n<p>The <strong>Sony-TAu Realistic Spatial Soundscapes 2023&nbsp;(STARSS23)</strong>&nbsp;dataset contains multichannel recordings of sound scenes in various rooms and environments, together with temporal and spatial annotations of prominent events belonging to a set of target classes. The dataset is collected in two different countries, in Tampere, Finland by the Audio Researh Group (ARG) of <strong>Tampere University (TAU)</strong>, and in Tokyo, Japan by <strong>SONY</strong>, using a similar setup and annotation procedure. The dataset is delivered in two 4-channel spatial recording formats, a microphone array one (<strong>MIC</strong>), and first-order Ambisonics one (<strong>FOA</strong>). These recordings serve as the development dataset for the&nbsp;<a href=\"https://dcase.community/challenge2023/task-sound-event-localization-and-detection\">DCASE 2023 Sound Event Localization and Detection Task</a>&nbsp;of the&nbsp;<a href=\"https://dcase.community/challenge2023/\">DCASE 2023 Challenge</a>.<br>\n<br>\nThe STARSS23 dataset is a continuation of the <a href=\"https://zenodo.org/record/6600531\">STARSS22 dataset</a>. It extends the previous version with the following:</p>\n\n<ul>\n\t<li>An additional&nbsp;<strong>2.5hrs&nbsp;</strong>of recordings in the development set, from&nbsp;<strong>5 new rooms</strong>&nbsp;distributed in 47 new recording clips.</li>\n\t<li><strong>Distance labels</strong> (in cm) for the spatially annotated sound events, apart from only the previous azimuth and elevation labels.</li>\n\t<li><strong>360&deg; videos</strong> spatially and temporally aligned to the audio recordings of the dataset (apart from 12 audio-only clips).</li>\n\t<li>Additional new audio and video recordings will be added in the evaluation set of the dataset in a subsequent version.</li>\n</ul>\n\n<p>Contrary to the three previous datasets of synthetic spatial sound scenes of TAU Spatial Sound Events 2019 (<a href=\"https://zenodo.org/record/2599196\">development</a>/<a href=\"https://zenodo.org/record/3377088\">evaluation</a>),&nbsp;<a href=\"https://doi.org/10.5281/zenodo.4064792\">TAU-NIGENS Spatial Sound Events 2020</a>, and&nbsp;<a href=\"https://zenodo.org/record/5476980\">TAU-NIGENS Spatial Sound Events 2021</a>&nbsp;associated with previous iterations of the DCASE Challenge, the STARS22-23 dataset contains recordings of real sound scenes and hence it avoids some of the pitfalls of synthetic generation of scenes. Some such key properties are:</p>\n\n<ul>\n\t<li>annotations are based on a combination of human annotators for sound event activity and optical tracking for spatial positions,</li>\n\t<li>the annotated target event classes are determined by the composition of the real scenes,</li>\n\t<li>the density, polyphony, occurences and co-occurences of events and sound classes is not random, and it follows actions and interactions of participants in the real scenes.</li>\n</ul>\n\n<p>The first round of recordings was collected between September 2021 and January 2022. A second round of recordings was collected between&nbsp;November 2022 and February 2023.<br>\n<br>\nCollection of data from the TAU side has received funding from Google.</p>\n\n<p><strong>REPORT &amp; REFERENCE:</strong></p>\n\n<p>If you use this dataset you could cite this report on its design, capturing, and annotation process:</p>\n\n<p>Archontis Politis,&nbsp;Kazuki Shimada,&nbsp;Parthasaarathy Sudarsanam,&nbsp;Sharath Adavanne,&nbsp;Daniel Krause,&nbsp;Yuichiro Koyama,&nbsp;Naoya Takahashi,&nbsp;Shusuke Takahashi,&nbsp;Yuki Mitsufuji,&nbsp;Tuomas Virtanen (2022).&nbsp;<strong>STARSS22: A dataset of spatial recordings of real scenes with spatiotemporal annotations of sound events</strong>.&nbsp;In&nbsp;<em>Proceedings of the Detection and Classification of Acoustic Scenes and Events 2022 Workshop (DCASE2022)</em>, Nancy, France.</p>\n\n<p>found&nbsp;<a href=\"https://dcase.community/documents/workshop2022/proceedings/DCASE2022Workshop_Politis_51.pdf\">here</a>.</p>\n\n<p>A more detailed report on the properties of the new dataset and its audiovisual processing with a suitable baseline for DCASE2023 will be published soon.</p>\n\n<p><strong>AIM:</strong></p>\n\n<p>The STARSS22-23 dataset is suitable for training and evaluation of machine-listening models for sound event detection (SED), general sound source localization with diverse sounds or signal-of-interest localization, and joint sound-event-localization-and-detection (SELD). Additionally, the dataset can be used for evaluation of signal processing methods that do not necessarily rely on training, such as acoustic source localization methods and multiple-source acoustic tracking. The dataset allows evaluation of the performance and robustness of the aforementioned applications for diverse types of sounds, and under diverse acoustic conditions.</p>\n\n<p>Specifically the STARSS23 allows additionally evaluation of audiovisual processing methods, such as audiovisual source localization.</p>\n\n<p><strong>SPECIFICATIONS:</strong></p>\n\n<p>General:</p>\n\n<ul>\n\t<li>Recordings are taken in two different sites.</li>\n\t<li>Each recording clip is part of a recording session happening in a unique room.</li>\n\t<li>Groups of participants, sound making props, and scene scenarios are unique for each session (with a few exceptions).</li>\n\t<li>To achieve good variability and efficiency in the data, in terms of presence, density, movement, and/or spatial distribution of the sounds events, the scenes are loosely scripted.</li>\n\t<li>13 target classes are identified in the recordings and strongly annotated by humans.</li>\n\t<li>Spatial annotations for those active events are captured by an optical tracking system.</li>\n\t<li>Sound events out of the target classes are considered as interference.</li>\n\t<li>Occurrences of up to 3 simultaneous events are fairly common, while higher numbers of overlapping events (up to 5) can occur but are rare.</li>\n</ul>\n\n<p>Volume, duration, and data split:</p>\n\n<ul>\n\t<li>A total of 16 unique rooms captured in the recordings, 4 in Tokyo and 12 in Tampere (development set).</li>\n\t<li>70 recording clips of 30 sec ~ 5 min durations, with a total time of ~2hrs, captured in Tokyo (development dataset).</li>\n\t<li>98 recording clips of 40 sec ~ 9 min durations, with a total time of ~5.5hrs, captured in Tampere (development dataset).</li>\n\t<li>A training-testing split is provided for reporting results using the development dataset.</li>\n\t<li>40 recordings contributed by Sony for the training split, captured in 2 rooms (dev-train-sony).</li>\n\t<li>30 recordings contributed by Sony for the testing split, captured in 2 rooms (dev-test-sony).</li>\n\t<li>50 recordings contributed by TAU for the training split, captured in 7 rooms (dev-train-tau).</li>\n\t<li>48 recordings contributed by TAU for the testing split, captured in 5 rooms (dev-test-tau).</li>\n\t<li>About ~3.5hrs of additional recordings from both sites, captured in different rooms from the development set, will be released later as the evaluation set.&nbsp;</li>\n</ul>\n\n<p>Audio:</p>\n\n<ul>\n\t<li>Sampling rate: 24kHz.</li>\n\t<li>Two 4-channel 3-dimensional recording formats: first-order Ambisonics (FOA) and tetrahedral microphone array (MIC).</li>\n</ul>\n\n<p>Video:</p>\n\n<ul>\n\t<li>Video 360&deg; format: equirectangular</li>\n\t<li>Video resolution: 1920x960</li>\n\t<li>Video frames per second (fps): 29.97</li>\n\t<li>All audio recordings are accompanied by synchronised video recordings, apart from 12 audio recordings with missing videos (<em>fold3_room21_mix001.wav&nbsp;-&nbsp;fold3_room21_mix012.wav</em>)</li>\n</ul>\n\n<p>More detailed information on the dataset can be found in the included README file.</p>\n\n<p><strong>SOUND CLASSES:</strong></p>\n\n<p>13 target sound event classes are annotated. The classes follow loosely the&nbsp;<a href=\"https://research.google.com/audioset/ontology/index.html\">Audioset ontology</a>.</p>\n\n<p>&nbsp; 0.&nbsp;<strong>Female speech, woman speaking</strong><br>\n&nbsp; 1.&nbsp;<strong>Male speech, man speaking</strong><br>\n&nbsp; 2.&nbsp;<strong>Clapping</strong><br>\n&nbsp; 3.&nbsp;<strong>Telephone</strong><br>\n&nbsp; 4.&nbsp;<strong>Laughter</strong><br>\n&nbsp; 5.&nbsp;<strong>Domestic sounds</strong><br>\n&nbsp; 6.&nbsp;<strong>Walk, footsteps</strong><br>\n&nbsp; 7.&nbsp;<strong>Door, open or close</strong><br>\n&nbsp; 8.&nbsp;<strong>Music</strong><br>\n&nbsp; 9.&nbsp;<strong>Musical instrument</strong><br>\n&nbsp; 10.&nbsp;<strong>Water tap, faucet</strong><br>\n&nbsp; 11.&nbsp;<strong>Bell</strong><br>\n&nbsp; 12.&nbsp;<strong>Knock</strong></p>\n\n<p>The content of some of these classes corresponds to events of a limited range of Audioset-related subclasses. For more information see the README file.</p>\n\n<p><strong>EXAMPLE APPLICATION:</strong></p>\n\n<p>An implementation of a trainable model of a convolutional recurrent neural network, performing joint SELD, trained and evaluated with this dataset is provided&nbsp;<a href=\"https://github.com/sharathadavanne/seld-dcase2023\">here</a>. This&nbsp;implementation will serve as the baseline method for the audio-only track in the&nbsp;DCASE 2023 Sound Event Localization and Detection Task.<br>\n<br>\nA baseline for the audiovisual track of&nbsp;DCASE 2023 Sound Event Localization and Detection Task will be published soon and referenced here.</p>\n\n<p><strong>DEVELOPMENT AND EVALUATION:</strong></p>\n\n<p>The current version (Version 1.0) of the dataset includes only the 168 development audio/video&nbsp;recordings and labels, used by the participants of Task 3 of DCASE2023&nbsp;Challenge to train and validate their submitted systems. Version 1.1 will be including additionally the evaluation audio and video recordings without labels, for the evaluation phase of DCASE2023.</p>\n\n<p>If researchers wish to compare their system against the submissions of DCASE2023 Challenge, they will have directly comparable results if they use the evaluation data as their testing set.</p>\n\n<p><strong>DOWNLOAD INSTRUCTIONS:</strong></p>\n\n<p>The file&nbsp;<strong><em>foa_dev.zip</em></strong>, correspond to audio data of the&nbsp;<strong>FOA&nbsp;</strong>recording format.<br>\nThe file&nbsp;<strong><em>mic_dev.zip</em></strong>, correspond to audio data of the&nbsp;<strong>MIC</strong>&nbsp;recording format.</p>\n\n<p>The file&nbsp;<strong><em>video_dev.zip&nbsp;</em></strong>contains the common videos for both audio formats.<br>\nThe file&nbsp;<strong><em>metadata_dev.zip</em></strong>&nbsp;contains the common metadata for both audio formats.</p>\n\n<p>Download the zip files corresponding to the format of interest and use your favourite compression tool to unzip these zip files.</p>",
    "publication_date": "2023-03-08",
    "publisher": "Zenodo",
    "references": [
      {
        "reference": "Archontis Politis, Kazuki Shimada, Parthasaarathy Sudarsanam, Sharath Adavanne, Daniel Krause, Yuichiro Koyama, Naoya Takahashi, Shusuke Takahashi, Yuki Mitsufuji, Tuomas Virtanen (2022). STARSS22: A dataset of spatial recordings of real scenes with spatiotemporal annotations of sound events. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2022 Workshop (DCASE2022), Nancy, France."
      },
      {
        "reference": "Archontis Politis, Sharath Adavanne, Daniel Krause, Antoine Deleforge, Prerak Srivastava, Tuomas Virtanen (2021). A Dataset of Dynamic Reverberant Sound Scenes with Directional Interferers for Sound Event Localization and Detection.\u00a0 In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2021 Workshop (DCASE2021), Barcelona, Spain."
      },
      {
        "reference": "Archontis Politis, Sharath Adavanne, and Tuomas Virtanen (2020). A Dataset of Reverberant Spatial Sound Scenes with Moving Sources for Sound Event Localization and Detection. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2020 Workshop (DCASE2020), Tokyo, Japan."
      },
      {
        "reference": "Sharath Adavanne, Archontis Politis, and Tuomas Virtanen (2019). A Multi-room reverberant dataset for sound event localization and detection. Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019), New York, NY, USA."
      }
    ],
    "related_identifiers": [
      {
        "identifier": "10.5281/zenodo.6600531",
        "relation_type": {
          "id": "isnewversionof",
          "title": {
            "de": "Ist eine neue Version von",
            "en": "Is new version of"
          }
        },
        "resource_type": {
          "id": "dataset",
          "title": {
            "de": "Datensatz",
            "en": "Dataset"
          }
        },
        "scheme": "doi"
      },
      {
        "identifier": "10.5281/zenodo.5476980",
        "relation_type": {
          "id": "references",
          "title": {
            "de": "Referenziert",
            "en": "References"
          }
        },
        "resource_type": {
          "id": "dataset",
          "title": {
            "de": "Datensatz",
            "en": "Dataset"
          }
        },
        "scheme": "doi"
      },
      {
        "identifier": "10.5281/zenodo.4064792",
        "relation_type": {
          "id": "references",
          "title": {
            "de": "Referenziert",
            "en": "References"
          }
        },
        "resource_type": {
          "id": "dataset",
          "title": {
            "de": "Datensatz",
            "en": "Dataset"
          }
        },
        "scheme": "doi"
      },
      {
        "identifier": "10.5281/zenodo.2599196",
        "relation_type": {
          "id": "references",
          "title": {
            "de": "Referenziert",
            "en": "References"
          }
        },
        "resource_type": {
          "id": "dataset",
          "title": {
            "de": "Datensatz",
            "en": "Dataset"
          }
        },
        "scheme": "doi"
      },
      {
        "identifier": "10.5281/zenodo.3377088",
        "relation_type": {
          "id": "references",
          "title": {
            "de": "Referenziert",
            "en": "References"
          }
        },
        "resource_type": {
          "id": "dataset",
          "title": {
            "de": "Datensatz",
            "en": "Dataset"
          }
        },
        "scheme": "doi"
      },
      {
        "identifier": "https://github.com/sharathadavanne/seld-dcase2022",
        "relation_type": {
          "id": "references",
          "title": {
            "de": "Referenziert",
            "en": "References"
          }
        },
        "resource_type": {
          "id": "software",
          "title": {
            "de": "Software",
            "en": "Software"
          }
        },
        "scheme": "url"
      }
    ],
    "resource_type": {
      "id": "dataset",
      "title": {
        "de": "Datensatz",
        "en": "Dataset"
      }
    },
    "rights": [
      {
        "description": {
          "en": "A short and simple permissive license with conditions only requiring preservation of copyright and license notices. Licensed works, modifications, and larger works may be distributed under different terms and without source code."
        },
        "id": "mit",
        "props": {
          "scheme": "spdx",
          "url": "https://opensource.org/licenses/MIT"
        },
        "title": {
          "en": "MIT License"
        }
      }
    ],
    "subjects": [
      {
        "subject": "sound event detection"
      },
      {
        "subject": "sound source localization"
      },
      {
        "subject": "spatial audio"
      },
      {
        "subject": "360-degree video"
      },
      {
        "subject": "audiovisual processing"
      },
      {
        "subject": "acoustic scene analysis"
      },
      {
        "subject": "source distance estimation"
      },
      {
        "subject": "machine listening"
      },
      {
        "subject": "Ambisonics"
      }
    ],
    "title": "STARSS23: Sony-TAu Realistic Spatial Soundscapes 2023",
    "version": "1.0.0"
  },
  "parent": {
    "access": {
      "owned_by": {
        "user": "96758"
      }
    },
    "communities": {
      "default": "2d7be671-0ce3-4c1e-9ed3-cfc92571ad1a",
      "entries": [
        {
          "access": {
            "member_policy": "open",
            "members_visibility": "public",
            "record_submission_policy": "open",
            "review_policy": "open",
            "visibility": "public"
          },
          "children": {
            "allow": false
          },
          "created": "2019-11-25T09:04:41.826171+00:00",
          "custom_fields": {},
          "deletion_status": {
            "is_deleted": false,
            "status": "P"
          },
          "id": "2d7be671-0ce3-4c1e-9ed3-cfc92571ad1a",
          "links": {},
          "metadata": {
            "curation_policy": "",
            "description": "Datasets related to the DCASE challenges.",
            "page": "<p>The goal of this communty is to gather datasets related to the DCASE challenge series. More information available on <a href=\"http://dcase.community/\">DCASE website</a>.</p>\r\n",
            "title": "Detection and Classification of Acoustic Scenes and Events"
          },
          "revision_id": 0,
          "slug": "dcase",
          "updated": "2019-11-25T09:06:23.594750+00:00"
        }
      ],
      "ids": [
        "2d7be671-0ce3-4c1e-9ed3-cfc92571ad1a"
      ]
    },
    "id": "7709051",
    "pids": {
      "doi": {
        "client": "datacite",
        "identifier": "10.5281/zenodo.7709051",
        "provider": "datacite"
      }
    }
  },
  "pids": {
    "doi": {
      "client": "datacite",
      "identifier": "10.5281/zenodo.7709052",
      "provider": "datacite"
    },
    "oai": {
      "identifier": "oai:zenodo.org:7709052",
      "provider": "oai"
    }
  },
  "revision_id": 7,
  "stats": {
    "all_versions": {
      "data_volume": 74403649423428.0,
      "downloads": 32220,
      "unique_downloads": 18218,
      "unique_views": 12066,
      "views": 13514
    },
    "this_version": {
      "data_volume": 12841548922373.0,
      "downloads": 5639,
      "unique_downloads": 3596,
      "unique_views": 2851,
      "views": 3220
    }
  },
  "status": "published",
  "swh": {},
  "updated": "2023-06-26T13:33:06.256826+00:00",
  "versions": {
    "index": 1,
    "is_latest": false
  }
}