CNNs improve decoding of selective attention to speech in cochlear implant users.

Journal: Journal of neural engineering
Published Date:

Abstract

. Understanding speech in the presence of background noise such as other speech streams is a difficult problem for people with hearing impairment, and in particular for users of cochlear implants (CIs). To improve their listening experience, auditory attention decoding (AAD) aims to decode the target speaker of a listener from electroencephalography (EEG), and then use this information to steer an auditory prosthesis towards this speech signal. In normal-hearing individuals, deep neural networks (DNNs) have been shown to improve AAD compared to simpler linear models. We aim to demonstrate that DNNs can improve attention decoding in CI users too, which would make them the state-of-the-art candidate for a neuro-steered CI.. To this end, we first collected an EEG dataset on selective auditory attention from 25 bilateral CI users, and then implemented both a linear model as well as a convolutional neural network (CNN) for attention decoding. Moreover, we introduced a novel, objective CI-artifact removal strategy and evaluated its impact on decoding accuracy, alongside learnable speaker classification using a support vector machine (SVM).. The CNN outperformed the linear model across all decision window sizes from 1 to 60 s. Removing CI artifacts modestly improved the CNN's decoding accuracy. With SVM classification, the CNN decoder reached a peak mean decoding accuracy of 74% at the population level for a 60 s decision window.. These results demonstrate the superior potential of CNN-based decoding for neuro-steered CIs, which could improve speech perception of its users in cocktail party situations significantly.

Authors

  • Constantin Jehn
    Department Artificial Intelligence in Biomedical Engineering (AIBE) at the Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany.
  • Adrian Kossmann
    Ear Research Center Dresden (ERCD), University Hospital Carl Gustav Carus, Technische Universität Dresden, Fetscherstraße 74, 01307 Dresden, Germany.
  • Niki Katerina Vavatzanidis
    Ear Research Center Dresden (ERCD), University Hospital Carl Gustav Carus, Technische Universität Dresden, Fetscherstraße 74, 01307 Dresden, Germany.
  • Anja Hahne
    Ear Research Center Dresden (ERCD), University Hospital Carl Gustav Carus, Technische Universität Dresden, Fetscherstraße 74, 01307 Dresden, Germany.
  • Tobias Reichenbach
    Imperial College London.