SparrKULee: A Speech-Evoked Auditory Response Repository from KU Leuven, Containing the EEG of 85 Participants
Bernd Accou (),
Lies Bollens (),
Marlies Gillis,
Wendy Verheijen,
Hugo Van hamme and
Tom Francart ()
Additional contact information
Bernd Accou: Experimental Oto-Rhino-Laryngology (ExpORL), Department Neurosciences, KU Leuven, B-3001 Leuven, Belgium
Lies Bollens: Experimental Oto-Rhino-Laryngology (ExpORL), Department Neurosciences, KU Leuven, B-3001 Leuven, Belgium
Marlies Gillis: Experimental Oto-Rhino-Laryngology (ExpORL), Department Neurosciences, KU Leuven, B-3001 Leuven, Belgium
Wendy Verheijen: Experimental Oto-Rhino-Laryngology (ExpORL), Department Neurosciences, KU Leuven, B-3001 Leuven, Belgium
Hugo Van hamme: Processing Speech and Images (PSI), Department of Electrical Engineering (ESAT), KU Leuven, B-3001 Leuven, Belgium
Tom Francart: Experimental Oto-Rhino-Laryngology (ExpORL), Department Neurosciences, KU Leuven, B-3001 Leuven, Belgium
Data, 2024, vol. 9, issue 8, 1-18
Abstract:
Researchers investigating the neural mechanisms underlying speech perception often employ electroencephalography (EEG) to record brain activity while participants listen to spoken language. The high temporal resolution of EEG enables the study of neural responses to fast and dynamic speech signals. Previous studies have successfully extracted speech characteristics from EEG data and, conversely, predicted EEG activity from speech features. Machine learning techniques are generally employed to construct encoding and decoding models, which necessitate a substantial quantity of data. We present SparrKULee, a Speech-evoked Auditory Repository of EEG data, measured at KU Leuven, comprising 64-channel EEG recordings from 85 young individuals with normal hearing, each of whom listened to 90–150 min of natural speech. This dataset is more extensive than any currently available dataset in terms of both the number of participants and the quantity of data per participant. It is suitable for training larger machine learning models. We evaluate the dataset using linear and state-of-the-art non-linear models in a speech encoding/decoding and match/mismatch paradigm, providing benchmark scores for future research.
Keywords: auditory EEG; dataset; speech (search for similar items in EconPapers)
JEL-codes: C8 C80 C81 C82 C83 (search for similar items in EconPapers)
Date: 2024
References: View references in EconPapers View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2306-5729/9/8/94/pdf (application/pdf)
https://www.mdpi.com/2306-5729/9/8/94/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jdataj:v:9:y:2024:i:8:p:94-:d:1443705
Access Statistics for this article
Data is currently edited by Ms. Cecilia Yang
More articles in Data from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().