Portable Facial Expression System Based on EMG Sensors and Machine Learning Models.

Journal: Sensors (Basel, Switzerland)
Published Date:

Abstract

One of the biggest challenges of computers is collecting data from human behavior, such as interpreting human emotions. Traditionally, this process is carried out by computer vision or multichannel electroencephalograms. However, they comprise heavy computational resources, far from final users or where the dataset was made. On the other side, sensors can capture muscle reactions and respond on the spot, preserving information locally without using robust computers. Therefore, the research subject is the recognition of the six primary human emotions using electromyography sensors in a portable device. They are placed on specific facial muscles to detect happiness, anger, surprise, fear, sadness, and disgust. The experimental results showed that when working with the CortexM0 microcontroller, enough computational capabilities were achieved to store a deep learning model with a classification store of 92%. Furthermore, we demonstrate the necessity of collecting data from natural environments and how they need to be processed by a machine learning pipeline.

Authors

  • Paola A Sanipatín-Díaz
    SDAS Research Group, Hay Moulay Rachid, Ben Guerir 43150, Morocco.
  • Paul D Rosero-Montalvo
    Computer Science Department, IT University of Copenhagen, 2300 Copenhagen, Denmark.
  • Wilmar Hernandez
    Carrera de Ingenieria Electronica y Automatizacion, Facultad de Ingenieria y Ciencias Aplicadas, Universidad de Las Americas, Quito 170124, Ecuador.