A CrossMod-Transformer deep learning framework for multi-modal pain detection through EDA and ECG fusion.

Journal: Scientific reports
Published Date:

Abstract

Pain is a multifaceted phenomenon that significantly affects a large portion of the global population. Objective pain assessment is essential for developing effective management strategies, which in turn contribute to more efficient and responsive healthcare systems. However, accurately evaluating pain remains a complex challenge due to subtle physiological and behavioural indicators, individual-specific pain responses, and the need for continuous patient monitoring. Automatic pain assessment systems offer promising, technology-driven solutions to support and enhance various aspects of the pain evaluation process. Physiological indicators offer valuable insights into pain-related states and are generally less influenced by individual variability compared to behavioural modalities, such as facial expressions. Skin conductance, regulated by sweat gland activity, and the heart's electrical signals are both influenced by changes in the sympathetic nervous system. Biosignals, such as electrodermal activity (EDA) and electrocardiogram (ECG), can, therefore, objectively capture the body's physiological responses to painful stimuli. This paper proposes a novel multi-modal ensemble deep learning framework that combines electrodermal activity and electrocardiogram signals for automatic pain recognition. The proposed framework includes a uni-modal approach (FCN-ALSTM-Transformer) comprising a Fully Convolutional Network, Attention-based LSTM, and a Transformer block to integrate features extracted by these models. Additionally, a multi-modal approach (CrossMod-Transformer) is introduced, featuring a dedicated Transformer architecture that fuses electrodermal activity and electrocardiogram signals. Experimental evaluations were primarily conducted on the BioVid dataset, with further cross-dataset validation using the AI4PAIN 2025 dataset to assess the generalisability of the proposed method. Notably, the CrossMod-Transformer achieved an accuracy of 87.52% on Biovid and 75.83% on AI4PAIN, demonstrating strong performance across independent datasets and outperforming several state-of-the-art uni-modal and multi-modal methods. These results highlight the potential of the proposed framework to improve the reliability of automatic multi-modal pain recognition and support the development of more objective and inclusive clinical assessment tools.

Authors

  • Jaleh Farmani
    University of Rome 'La Sapienza', Department of Computer, Control & Management Engineering, Rome, 00185, Italy.
  • Ghazal Bargshady
    School of Management and Enterprise, University of Southern Queensland, Springfield, Qld 4300, Australia. Electronic address: ghazal.bargshady@usq.edu.au.
  • Stefanos Gkikas
    Department of Electrical and Computer Engineering, Hellenic Mediterranean University, Estavromenos, Heraklion, 71410, Greece; Computational BioMedicine Laboratory, Institute of Computer Science, Foundation for Research & Technology-Hellas, Vassilika Vouton, Heraklion, 70013, Greece. Electronic address: gkikas@ics.forth.gr.
  • Manolis Tsiknakis
    Computational BioMedicine Laboratory, FORTH-ICS, Heraklion, Crete, Greece.
  • Raul Fernandez Rojas
    University of Canberra, Faculty of Science & Technology, Canberra, 2917, Australia. raul.fernandezrojas@canberra.edu.au.

Keywords

No keywords available for this article.