Dataset Open Access

EEG and audio dataset for auditory attention decoding

Fuglsang, Søren A.; Wong, Daniel D.E.; Hjortkjær, Jens

Citation Style Language JSON Export

  "publisher": "Zenodo", 
  "DOI": "10.5281/zenodo.1199011", 
  "title": "EEG and audio dataset for auditory attention decoding", 
  "issued": {
    "date-parts": [
  "abstract": "<p>This dataset contains EEG recordings from 18 subjects listening to one of two competing speech audio streams. Continuous speech in trials of ~50 sec. was presented to normal hearing listeners in simulated rooms with different degrees of reverberation. Subjects were asked to attend one of two spatially separated speakers (one male, one female) and ignore the other. Repeated trials with presentation of a single talker were also recorded. The data were recorded in a double-walled soundproof booth at the Technical University of Denmark (DTU) using a 64-channel Biosemi system and digitized at a sampling rate of 512 Hz. Full details can be found in:</p>\n\n<ul>\n\t<li><strong>S&oslash;ren A. Fuglsang, Torsten Dau &amp; Jens Hjortkj&aelig;r (2017):&nbsp;Noise-robust cortical tracking of attended speech in real-life environments. <em>NeuroImage</em>, 156, 435-444</strong></li>\n</ul>\n\n<p>and</p>\n\n<ul>\n\t<li><strong>Daniel D.E. Wong, S&oslash;ren A. Fuglsang, Jens Hjortkj&aelig;r, Enea Ceolini, Malcolm Slaney &amp; Alain de Cheveign&eacute;: A Comparison of Temporal Response Function Estimation Methods for Auditory Attention Decoding. Frontiers in Neuroscience,&nbsp;</strong><a href=\"\"></a></li>\n</ul>\n\n<p>The data is organized in format of the publicly available <a href=\"\">COCOHA Matlab Toolbox</a>. The preproc_script.m demonstrates how to import and align the EEG and audio data. The script also demonstrates some EEG preprocessing steps as used the Wong et al. paper above. The contains wav-files with the speech audio used in the experiment. The contains MAT-files with the EEG/EOG data for each subject. The EEG/EOG data are found in <strong>data.eeg</strong> with the following channels:</p>\n\n<ul>\n\t<li>channels 1-64: scalp EEG electrodes</li>\n\t<li>channel 65: right mastoid electrode</li>\n\t<li>channel 66: left mastoid electrode</li>\n\t<li>channel 67: vertical EOG below right eye</li>\n\t<li>channel 68: horizontal EOG right eye</li>\n\t<li>channel 69: vertical EOG above right eye</li>\n\t<li>channel 70: vertical EOG below left eye</li>\n\t<li>channel 71: horizontal EOG left eye</li>\n\t<li>channel 72: vertical EOG above left eye</li>\n</ul>\n\n<p>The <strong>expinfo</strong> table contains information about experimental conditions, including what what speaker the listener was attending to in different trials. The expinfo table contains the following information:</p>\n\n<ul>\n\t<li>attend_mf: attended speaker (1=male, 2=female)</li>\n\t<li>attend_lr: spatial position of the attended speaker (1=left, 2=right)</li>\n\t<li>acoustic_condition: type of acoustic room (1= anechoic, 2= mild reverberation, 3= high reverberation, see Fuglsang et al. for details)</li>\n\t<li>n_speakers: number of speakers presented (1 or 2)</li>\n\t<li>wavfile_male: name of presented audio wav-file for the male speaker</li>\n\t<li>wavfile_female: name of presented audio wav-file for the female speaker (if any)</li>\n\t<li>trigger: trigger event value for each trial also found in data.event.eeg.value</li>\n</ul>\n\n<p> contains the preprocessed EEG and audio data as output from preproc_script.m.</p>\n\n<p>The dataset was created within the <a href=\"\">COCOHA</a><a href=\"\"> Project</a>: Cognitive Control of a Hearing Aid</p>", 
  "author": [
      "family": "Fuglsang, S\u00f8ren A."
      "family": "Wong, Daniel D.E."
      "family": "Hjortkj\u00e6r, Jens"
  "version": "1", 
  "type": "dataset", 
  "id": "1199011"
All versions This version
Views 3,4833,486
Downloads 2,5932,593
Data volume 24.8 TB24.8 TB
Unique views 3,1593,162
Unique downloads 1,0751,075


Cite as