AIMC Topic: Speech

Clear Filters Showing 41 to 50 of 352 articles

Multimodal machine learning for language and speech markers identification in mental health.

BMC medical informatics and decision making
BACKGROUND: There are numerous papers focusing on diagnosing mental health disorders using unimodal and multimodal approaches. However, our literature review shows that the majority of these studies either use unimodal approaches to diagnose a variet...

Detection of hate: speech tweets based convolutional neural network and machine learning algorithms.

Scientific reports
There is no doubt that social media sites have provided many benefits to humanity, such as sharing information continuously and communicating with others easily. It also seems that social media sites have many advantages, but in addition to these adv...

Deep temporal representation learning for language identification.

Neural networks : the official journal of the International Neural Network Society
Language identification (LID) is a key component in downstream tasks. Recently, the self-supervised speech representation learned by Wav2Vec 2.0 (W2V2) has been demonstrated to be very effective for various speech-related tasks. In LID, it is commonl...

Roman urdu hate speech detection using hybrid machine learning models and hyperparameter optimization.

Scientific reports
With the rapid increase of users over social media, cyberbullying, and hate speech problems have arisen over the past years. Automatic hate speech detection (HSD) from text is an emerging research problem in natural language processing (NLP). Researc...

The voice of depression: speech features as biomarkers for major depressive disorder.

BMC psychiatry
BACKGROUND: Psychiatry faces a challenge due to the lack of objective biomarkers, as current assessments are based on subjective evaluations. Automated speech analysis shows promise in detecting symptom severity in depressed patients. This project ai...

Speech production under stress for machine learning: multimodal dataset of 79 cases and 8 signals.

Scientific data
Early identification of cognitive or physical overload is critical in fields where human decision making matters when preventing threats to safety and property. Pilots, drivers, surgeons, and operators of nuclear plants are among those affected by th...

Deep-learning models reveal how context and listener attention shape electrophysiological correlates of speech-to-language transformation.

PLoS computational biology
To transform continuous speech into words, the human brain must resolve variability across utterances in intonation, speech rate, volume, accents and so on. A promising approach to explaining this process has been to model electroencephalogram (EEG) ...

Speech based suicide risk recognition for crisis intervention hotlines using explainable multi-task learning.

Journal of affective disorders
BACKGROUND: Crisis Intervention Hotline can effectively reduce suicide risk, but suffer from low connectivity rates and untimely crisis response. By integrating speech signals and deep learning to assist in crisis assessment, it is expected to enhanc...

Speech recognition using an english multimodal corpus with integrated image and depth information.

Scientific reports
Traditional English corpora mainly collect information from a single modality, but lack information from multimodal information, resulting in low quality of corpus information and certain problems with recognition accuracy. To solve the above problem...

Federated learning and deep learning framework for MRI image and speech signal-based multi-modal depression detection.

Computational biology and chemistry
Adolescence is a significant period for developing skills and knowledge and learning about managing relationships and emotions by gathering attributes for maturity. Recently, Depression arises as a common mental health issue in adolescents and this a...