Dataset Open Access

VocalSketch Data Set v1.0.4

Cartwright, Mark; Pardo, Bryan


MARC21 XML Export

<?xml version='1.0' encoding='UTF-8'?>
<record xmlns="http://www.loc.gov/MARC21/slim">
  <leader>00000nmm##2200000uu#4500</leader>
  <datafield tag="999" ind1="C" ind2="5">
    <subfield code="x">Cartwright, M., Pardo, B. VocalSketch: Vocally Imitating Audio Concepts. In Proceedings of ACM Conference on Human Factors in Computing Systems (2015). http://dx.doi.org/10.1145/2702123.2702387</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">user interaction</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">vocal imitation</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">audio software</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">foley</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">audio search</subfield>
  </datafield>
  <datafield tag="653" ind1=" " ind2=" ">
    <subfield code="a">sound</subfield>
  </datafield>
  <controlfield tag="005">20170906064832.0</controlfield>
  <controlfield tag="001">13862</controlfield>
  <datafield tag="711" ind1=" " ind2=" ">
    <subfield code="d">18-23 April 2015</subfield>
    <subfield code="g">CHI</subfield>
    <subfield code="a">ACM Conference on Human Factors in Computing Systems</subfield>
    <subfield code="c">Seoul, Korea</subfield>
  </datafield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="u">Interactive Audio Lab, Department of Electrical Engineering and Computer Science, Northwestern University</subfield>
    <subfield code="a">Pardo, Bryan</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2=" ">
    <subfield code="s">1991818509</subfield>
    <subfield code="z">md5:d719085b90a8fa28c18fad897e1719ad</subfield>
    <subfield code="u">https://zenodo.org/record/13862/files/VocalSketchDataSet-1.0.4.rar</subfield>
  </datafield>
  <datafield tag="542" ind1=" " ind2=" ">
    <subfield code="l">open</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2=" ">
    <subfield code="y">Conference website</subfield>
    <subfield code="u">http://chi2015.acm.org/</subfield>
  </datafield>
  <datafield tag="260" ind1=" " ind2=" ">
    <subfield code="c">2015-01-09</subfield>
  </datafield>
  <datafield tag="909" ind1="C" ind2="O">
    <subfield code="p">openaire_data</subfield>
    <subfield code="o">oai:zenodo.org:13862</subfield>
  </datafield>
  <datafield tag="100" ind1=" " ind2=" ">
    <subfield code="u">Interactive Audio Lab, Department of Electrical Engineering and Computer Science, Northwestern University</subfield>
    <subfield code="a">Cartwright, Mark</subfield>
  </datafield>
  <datafield tag="245" ind1=" " ind2=" ">
    <subfield code="a">VocalSketch Data Set v1.0.4</subfield>
  </datafield>
  <datafield tag="540" ind1=" " ind2=" ">
    <subfield code="u">https://creativecommons.org/licenses/by/4.0/</subfield>
    <subfield code="a">Creative Commons Attribution 4.0</subfield>
  </datafield>
  <datafield tag="650" ind1="1" ind2="7">
    <subfield code="a">cc-by</subfield>
    <subfield code="2">opendefinition.org</subfield>
  </datafield>
  <datafield tag="520" ind1=" " ind2=" ">
    <subfield code="a">&lt;p&gt;This data set contains thousands of vocal imitations of a large set of diverse sounds. These imitations were collected from hundreds of contributors via Amazon&amp;#39;s Mechanical Turk website. The data set also contains data on hundreds of people&amp;#39;s ability to correctly label these vocal imitations, also collected via Amazon&amp;#39;s Mechanical Turk. This data set will help the research community understand which audio concepts can be effectively communicated with this approach. We have released this data so the community can study the related issues and build systems that leverage vocal imitation as an interaction modality, such as search engines that can be queried by vocally imitating the desired sound.&lt;/p&gt;

&lt;p&gt;&amp;nbsp;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;This data set is a supplement to a paper. Please cite the following paper to reference this data set in a publication:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Cartwright, M., Pardo, B. VocalSketch: Vocally Imitating Audio Concepts. In &lt;em&gt;Proceedings of ACM Conference on Human Factors in Computing Systems&lt;/em&gt;&amp;nbsp;(2015). http://dx.doi.org/10.1145/2702123.2702387&lt;/p&gt;

&lt;p&gt;&amp;nbsp;&lt;/p&gt;

&lt;p&gt;See&amp;nbsp;https://github.com/interactiveaudiolab/VocalSketchDataSet for the latest updates to this data set.&lt;/p&gt;

&lt;p&gt;&amp;nbsp;&lt;/p&gt;

&lt;p&gt;Interactive Audio Lab: http://music.eecs.northwestern.edu&lt;/p&gt;</subfield>
  </datafield>
  <datafield tag="773" ind1=" " ind2=" ">
    <subfield code="n">doi</subfield>
    <subfield code="i">isSupplementTo</subfield>
    <subfield code="a">10.1145/2702123.2702387</subfield>
  </datafield>
  <datafield tag="773" ind1=" " ind2=" ">
    <subfield code="n">url</subfield>
    <subfield code="i">isIdenticalTo</subfield>
    <subfield code="a">https://github.com/interactiveaudiolab/VocalSketchDataSet/releases/tag/v1.0.4</subfield>
  </datafield>
  <datafield tag="024" ind1=" " ind2=" ">
    <subfield code="a">10.5281/zenodo.13862</subfield>
    <subfield code="2">doi</subfield>
  </datafield>
  <datafield tag="980" ind1=" " ind2=" ">
    <subfield code="a">dataset</subfield>
  </datafield>
</record>

Share

Cite as