EEG-based speech imagery decoding by dynamic hypergraph learning within projected and selected feature subspaces.

Journal: Journal of neural engineering
Published Date:

Abstract

Speech imagery is a nascent paradigm that is receiving widespread attention in current brain-computer interface (BCI) research. By collecting the electroencephalogram (EEG) data generated when imagining the pronunciation of a sentence or word in human mind, machine learning methods are used to decode the intention that the subject wants to express. Among existing decoding methods, graph is often used as an effective tool to model the data structure; however, in the field of BCI research, the correlations between EEG samples may not be fully characterized by simple pairwise relationships. Therefore, this paper attempts to employ a more effective data structure to model EEG data.In this paper, we introduce hypergraph to describe the high-order correlations between samples by viewing feature vectors extracted from each sample as vertices and then connecting them through hyperedges. We also dynamically update the weights of hyperedges, the weights of vertices and the structure of the hypergraph in two transformed subspaces, i.e. projected and feature-weighted subspaces. Accordingly, two dynamic hypergraph learning models, i.e. dynamic hypergraph semi-supervised learning within projected subspace (DHSLP) and dynamic hypergraph semi-supervised learning within selected feature subspace (DHSLF), are proposed for speech imagery decoding.To validate the proposed models, we performed a series of experiments on two EEG datasets. The obtained results demonstrated that both DHSLP and DHSLF have statistically significant improvements in decoding imagined speech intentions to existing studies. Specifically, DHSLP achieved accuracies of 78.40% and 66.64% on the two datasets, while DHSLF achieved accuracies of 71.07% and 63.94%.Our study indicates the effectiveness of the learned hypergraphs in characterizing the underlying semantic information of imagined contents; besides, interpretable results on quantitatively exploring the discriminative EEG channels in speech imagery decoding are obtained, which lay the foundation for further exploration of the physiological mechanisms during speech imagery.

Authors

  • Yibing Li
    School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, People's Republic of China.
  • Zhenye Zhao
    School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, People's Republic of China.
  • Jiangchuan Liu
    HDU-ITMO Joint Institute, Hangzhou Dianzi University, Hangzhou 310018, People's Republic of China.
  • Yong Peng
    Department of Cardiology, West China Hospital, Sichuan University, Chengdu 610041, China.
  • Kenneth Camilleri
    Centre for Biomedical Cybernetics, University of Malta, 2080 Misa, Malta.
  • Wanzeng Kong
    School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, Zhejiang 310018, China.
  • Andrzej Cichocki