Decoding selective auditory attention with EEG using a transformer model

•Attention-mechanism is effective in improving the performance of EEG-based speech envelope reconstruction.•An AAD-transformer including temporal self-attention and channel attention modules is proposed for AAD.•AAD-transformer is data-driven with no requirement of preprocessing of EEG. The human au...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Methods (San Diego, Calif.) Calif.), 2022-08, Vol.204, p.410-417
Hauptverfasser: Xu, Zihao, Bai, Yanru, Zhao, Ran, Hu, Hongmei, Ni, Guangjian, Ming, Dong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•Attention-mechanism is effective in improving the performance of EEG-based speech envelope reconstruction.•An AAD-transformer including temporal self-attention and channel attention modules is proposed for AAD.•AAD-transformer is data-driven with no requirement of preprocessing of EEG. The human auditory system extracts valid information in noisy environments while ignoring other distractions, relying primarily on auditory attention. Studies have shown that the cerebral cortex responds differently to the sound source locations and that auditory attention is time-varying. In this work, we proposed a data-driven encoder-decoder architecture model for auditory attention detection (AAD), denoted as AAD-transformer. The model contains temporal self-attention and channel attention modules and could reconstruct the speech envelope by dynamically assigning weights according to the temporal self-attention and channel attention mechanisms of electroencephalogram (EEG). In addition, the model is conducted based on data-driven without additional preprocessing steps. The proposed model was validated using a binaural listening dataset, in which the speech stimulus was Mandarin, and compared with other models. The results showed that the decoding accuracy of the AAD-transformer in the 0.15-second decoding time window was 76.35%, which was much higher than the accuracy of the linear model using temporal response function in the 3-second decoding time window (increased by 16.27%). This work provides a novel auditory attention detection method, and the data-driven characteristic makes it convenient for neural-steered hearing devices, especially those who speak tonal languages.
ISSN:1046-2023
1095-9130
DOI:10.1016/j.ymeth.2022.04.009