Abstract
Previous studies on neurosteered hearing aids employed a neural decoder to reconstruct the speech stimulus from electroencephalogram (EEG) signals to establish the attended sound source. However, this approach presents several limitations, such as the need for clean speech stimuli—which are often unavailable in real-world scenarios—and long processing windows. To address these challenges, we propose a novel EEG-based neurosteered speaker extraction (ENSE) mechanism that performs a joint action of speech separation and direct attention classification without the need for explicit speech stimulus reconstruction. Specifically, a typical speech separation model is first pretrained on a large speech corpus. We then train a speech-EEG match detector to perform direct attention classification by detecting which of the separated speech stimuli, or which of the speakers, induces the observed EEG signals. Experimental results show that ENSE effectively identifies and extracts the attended speech while suppressing unattended ones in a mixture. With time-domain speech separation and direct attention classification, ENSE offers a low-latency solution that marks an important step towards practical neurosteered hearing prostheses.
| Original language | English |
|---|---|
| Pages (from-to) | 102-112 |
| Number of pages | 11 |
| Journal | IEEE Transactions on Cognitive and Developmental Systems |
| Volume | 18 |
| Issue number | 1 |
| DOIs | |
| State | Published - Feb 2026 |
| Externally published | Yes |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Keywords
- Auditory attention detection (AAD)
- electroencephalogram
- hearing aid
- speech separation
Fingerprint
Dive into the research topics of 'EEG-Based Neurosteered Speaker Extraction in Cocktail Party Environment Without Stimulus Reconstruction'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver