Multimodal Depression Recognition via Mutual Information Maximization Joint with Multi-task Learning.
Journal:
IEEE transactions on bio-medical engineering
Published Date:
Jun 24, 2025
Abstract
Depression is a serious mental health disorder with a potential hazard for individuals and society characterized by persistent sadness and hopelessness. Multimodal information including vision, audio, and text is critical for depression diagnosis and treatment. Most studies focus on designing sophisticated feature extraction methods but ignore feature enhancement and fusion within intra-modality and cross-modality. In this paper, a Chinese Multimodal Depression Corpus (CMD-Corpus) dataset is established assisted by clinical experts aiming to support more depression research. Furthermore, we propose a multimodal depression recognition framework based on Mutual Information Maximization with Multi-task Learning (MIMML) to enhance feature representation and fusion among video, audio, and text modalities. The MIMML employs the strategy of maximizing mutual information to accelerate modality-invariance enhancement. The multi-task is used to improve the representation performance of the single modality to improve modality-specific enhancement. Meanwhile, a gated structure with bidirectional gated recurrent units and convolutional neural networks is designed to achieve multimodal feature fusion, which is key to boosting completeness among modalities. Experimental results show that the proposed MIMML effectively captures representation to increase depression recognition accuracy, achieving 84% and 89% accuracy on DAIC-WOZ and our self-collected CMD-Corpus dataset respectively.
Authors
Keywords
No keywords available for this article.