Artificial intelligence based multimodal language decoding from brain activity: A review.

Journal: Brain research bulletin
Published Date:

Abstract

Decoding brain activity is conducive to the breakthrough of brain-computer interface (BCI) technology. The development of artificial intelligence (AI) continually promotes the progress of brain language decoding technology. Existent research has mainly focused on a single modality and paid insufficient attention to AI methods. Therefore, our objective is to provide an overview of relevant decoding research from the perspective of different modalities and methodologies. The modalities involve text, speech, image, and video, whereas the core method is using AI-built decoders to translate brain signals induced by multimodal stimuli into text or vocal language. The semantic information of brain activity can be successfully decoded into a language at various levels, ranging from words through sentences to discourses. However, the decoding effect is affected by various factors, such as the decoding model, vector representation model, and brain regions. Challenges and future directions are also discussed. The advances in brain language decoding and BCI technology will potentially assist patients with clinical aphasia in regaining the ability to communicate.

Authors

  • Yuhao Zhao
    College of Language Intelligence, Sichuan International Studies University, Chongqing 400031, PR China.
  • Yu Chen
    State Key Laboratory of Oral Diseases & National Center for Stomatology & National Clinical Center for Oral Diseases, West China Hospital of Stomatology, Sichuan University, Chengdu, Sichuan, China.
  • Kaiwen Cheng
    School of Language Intelligence, Sichuan International Studies University, Chongqing, China.
  • Wei Huang
    Shaanxi Institute of Flexible Electronics, Northwestern Polytechnical University, 710072 Xi'an, China.