Using Ear-EEG to Decode Auditory Attention in Multiple-speaker Environment

  • Haolin Zhu
  • , Yujie Yan
  • , Xiran Xu
  • , Zhongshu Ge
  • , Pei Tian
  • , Xihong Wu
  • , Jing Chen

Research output: Contribution to journalConference articlepeer-review

2 Citations (Scopus)

Abstract

Auditory Attention Decoding (AAD) can help to determine the identity of the attended speaker during an auditory selective attention task, by analyzing and processing measurements of electroencephalography (EEG) data. Most studies on AAD are based on scalp-EEG signals in two-speaker scenarios, which are far from real application. Ear-EEG has recently gained significant attention due to its motion tolerance and invisibility during data acquisition, making it easy to incorporate with other devices for applications. In this work, participants selectively attended to one of the four spatially separated speakers’ speech in an anechoic room. The EEG data were concurrently collected from a scalp-EEG system and an ear-EEG system (cEEGrids). Temporal response functions (TRFs) and stimulus reconstruction (SR) were utilized using ear-EEG data. Results showed that the attended speech TRFs were stronger than each unattended speech and decoding accuracy was 41.3% in the 60s (chance level of 25%). To further investigate the impact of electrode placement and quantity, SR was utilized in both scalp-EEG and ear-EEG, revealing that while the number of electrodes had a minor effect, their positioning had a significant influence on the decoding accuracy. One kind of auditory spatial attention detection (ASAD) method, STAnet, was testified with this ear-EEG database, resulting in 93.1% in 1-second decoding window. The implementation code and database for our work are available on GitHub: https://github.com/zhl486/Ear_EEG_code.git and Zenodo: https://zenodo.org/records/10803261.

Keywords

  • auditory attention decoding
  • cEEGrids
  • cocktail party problem
  • ear-EEG
  • stimulus reconstruction

Fingerprint

Dive into the research topics of 'Using Ear-EEG to Decode Auditory Attention in Multiple-speaker Environment'. Together they form a unique fingerprint.

Cite this