CNNs improve decoding of selective attention to speech in cochlear implant users.
Journal:
Journal of neural engineering
Published Date:
Jun 10, 2025
Abstract
. Understanding speech in the presence of background noise such as other speech streams is a difficult problem for people with hearing impairment, and in particular for users of cochlear implants (CIs). To improve their listening experience, auditory attention decoding (AAD) aims to decode the target speaker of a listener from electroencephalography (EEG), and then use this information to steer an auditory prosthesis towards this speech signal. In normal-hearing individuals, deep neural networks (DNNs) have been shown to improve AAD compared to simpler linear models. We aim to demonstrate that DNNs can improve attention decoding in CI users too, which would make them the state-of-the-art candidate for a neuro-steered CI.. To this end, we first collected an EEG dataset on selective auditory attention from 25 bilateral CI users, and then implemented both a linear model as well as a convolutional neural network (CNN) for attention decoding. Moreover, we introduced a novel, objective CI-artifact removal strategy and evaluated its impact on decoding accuracy, alongside learnable speaker classification using a support vector machine (SVM).. The CNN outperformed the linear model across all decision window sizes from 1 to 60 s. Removing CI artifacts modestly improved the CNN's decoding accuracy. With SVM classification, the CNN decoder reached a peak mean decoding accuracy of 74% at the population level for a 60 s decision window.. These results demonstrate the superior potential of CNN-based decoding for neuro-steered CIs, which could improve speech perception of its users in cocktail party situations significantly.