There is a newer version of the record available.

Published September 21, 2022 | Version 1.0
Dataset Open

ESAA: an EEG-Speech auditory attention detection database

  • 1. National University of Singapore
  • 2. South China University of Technology
  • 3. The Chinese University of Hong Kong, Shenzhen

Description

We build a database for AAD research, which consists of competing speech stimuli and associated human neural responses, i.e, electroencephalography (EEG) recordings, namely EEG-Speech AAD (ESAA) database.  This is the first AAD database with speech stimuli in a tonal language (Mandarin). Moreover, we develop an AAD baseline as a reference model for decoding which speech stream a listening subject is attending to (speaker attention detection), and a baseline for decoding which spatial locus a listening subject is attending to (speaker locus attention detection) on the ESAA database.

We release the source code and the database for use in research purpose.

This database consists of response data for 17 normal-hearing subjects (S1-S17). It includes:

- 64-channel EEG data: responses to two-speaker speech stimuli
- Auditory stimuli data (clean): Chinese short stories narrated by a female and a male professional story teller. 
- Auditory stimuli data (hrtf): Auditory stimuli after head-related transfer function (HRTF) filtering (simulating sound coming from ± 90 deg).
- Preprocessing code
- AAD baseline (CNN model)

Notes

The work is funded by the National Natural Science Foundation of China (Grant No. 52075177). This research project is supported by IAF, A*STAR, SOITEC, NXP and National University of Singapore under FD-fAbrICS: Joint Lab for FD-SOI Always-on Intelligent $\&$ Connected Systems (Award I2001E0053). Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not reflect the views of the institutions and companies supporting the joint lab. The work by Haizhou Li is also supported by the Guangdong Provincial Key Laboratory of Big Data Computing, The Chinese University of Hong Kong, Shenzhen (Grant No. B10120210117-KP02).

Files

clean.zip

Files (13.7 GB)

Name Size Download all
md5:3cd1a57fcc3b5e4a399f9dc0afd1b1b5
404.5 MB Preview Download
md5:ed5fefc3661767adb7218b09602cfdd7
24.3 kB Preview Download
md5:9a73df6005bf827cc040dd5f06a86e5f
406.2 MB Preview Download
md5:007d20255ecf1687fbe1906d3e6fe344
17.9 kB Preview Download
md5:4d97f35c96b67a06dc746895928b55f3
974 Bytes Preview Download
md5:9c31e58d63fd0c816fbe919e850f54e3
528.1 MB Preview Download
md5:2346cfd81f9f6616bae0cc2652c753a7
646.5 MB Preview Download
md5:238066ea05b86bb2e8c1a9ff6ee385e3
841.7 MB Preview Download
md5:fbe1c5a877374cd0b95aab5d90ff3c62
814.3 MB Preview Download
md5:2ea0de7ae0b5bea9f1e7da789ad26a97
804.1 MB Preview Download
md5:dfb9f95dbb99e85378c70898f91ad3bf
808.7 MB Preview Download
md5:c91a3f998956c7642eb81493be223ea5
809.8 MB Preview Download
md5:a62588dc0a1505c96512378f5a393cf8
634.2 MB Preview Download
md5:66a029133212c706c1d9a2b772c628d4
651.7 MB Preview Download
md5:223829f023fb672121ae706d21fc32ed
682.1 MB Preview Download
md5:15022f846f948d6e4085192c6d6eb59f
842.6 MB Preview Download
md5:c5947714eea8097247cccdb120d13b6e
677.1 MB Preview Download
md5:98b469e88ff2b48628cc718cf21e5b2c
667.8 MB Preview Download
md5:66b609e186eff5133747bb087d7695fd
829.3 MB Preview Download
md5:c920b1082437956922f194a2746dd5c3
770.2 MB Preview Download
md5:08ecd9ac416bcf5774043d77a49cf1f2
1.2 GB Preview Download
md5:a3b763ea578770c29ece92019420aba5
693.0 MB Preview Download

Additional details

References

  • Please refer the following paper if you use this database
  • P. Li, E. Su, J. Li, S. Cai, L. Xie and H. Li, "ESAA: An Eeg-Speech Auditory Attention Detection Database," 2022 25th Conference of the Oriental COCOSDA International Committee for the Co-ordination and Standardisation of Speech Databases and Assessment Techniques (O-COCOSDA), Hanoi, Vietnam, 2022, pp. 1-6, doi: 10.1109/O-COCOSDA202257103.2022.9997944.
  • Cai, S., Su, E., Xie, L., & Li, H. (2021). EEG-Based Auditory Attention Detection via Frequency and Channel Neural Attention. IEEE Transactions on Human-Machine Systems, 52(2), 256-266.
  • Su, E., Cai, S., Xie, L., Li, H., & Schultz, T. (2022). STAnet: A spatiotemporal attention network for decoding auditory spatial attention from EEG. IEEE Transactions on Biomedical Engineering, 69(7), 2233-2242.