AIMC Topic: Speech

Clear Filters Showing 61 to 70 of 368 articles

Speech production under stress for machine learning: multimodal dataset of 79 cases and 8 signals.

Scientific data
Early identification of cognitive or physical overload is critical in fields where human decision making matters when preventing threats to safety and property. Pilots, drivers, surgeons, and operators of nuclear plants are among those affected by th...

Deep-learning models reveal how context and listener attention shape electrophysiological correlates of speech-to-language transformation.

PLoS computational biology
To transform continuous speech into words, the human brain must resolve variability across utterances in intonation, speech rate, volume, accents and so on. A promising approach to explaining this process has been to model electroencephalogram (EEG) ...

Speech based suicide risk recognition for crisis intervention hotlines using explainable multi-task learning.

Journal of affective disorders
BACKGROUND: Crisis Intervention Hotline can effectively reduce suicide risk, but suffer from low connectivity rates and untimely crisis response. By integrating speech signals and deep learning to assist in crisis assessment, it is expected to enhanc...

Speech recognition using an english multimodal corpus with integrated image and depth information.

Scientific reports
Traditional English corpora mainly collect information from a single modality, but lack information from multimodal information, resulting in low quality of corpus information and certain problems with recognition accuracy. To solve the above problem...

Federated learning and deep learning framework for MRI image and speech signal-based multi-modal depression detection.

Computational biology and chemistry
Adolescence is a significant period for developing skills and knowledge and learning about managing relationships and emotions by gathering attributes for maturity. Recently, Depression arises as a common mental health issue in adolescents and this a...

Human-AI collaboration improves adults' oral biomechanical functions: A multi-centre, self-controlled clinical trial.

Journal of dentistry
OBJECTIVES: Maintenance of oral muscle functions is important for survival and communication. Utilizing Artificial Intelligence (AI) as a self-health-management material has shown promise. Here we developed a functional and AI-enabled smartphone e-Or...

Effective Phoneme Decoding With Hyperbolic Neural Networks for High-Performance Speech BCIs.

IEEE transactions on neural systems and rehabilitation engineering : a publication of the IEEE Engineering in Medicine and Biology Society
OBJECTIVE: Speech brain-computer interfaces (speech BCIs), which convert brain signals into spoken words or sentences, have demonstrated great potential for high-performance BCI communication. Phonemes are the basic pronunciation units. For monosylla...

Machine learning-based classification of Parkinson's disease using acoustic features: Insights from multilingual speech tasks.

Computers in biology and medicine
This study advances the automation of Parkinson's disease (PD) diagnosis by analyzing speech characteristics, leveraging a comprehensive approach that integrates a voting-based machine learning model. Given the growing prevalence of PD, especially am...

A Combined CNN Architecture for Speech Emotion Recognition.

Sensors (Basel, Switzerland)
Emotion recognition through speech is a technique employed in various scenarios of Human-Computer Interaction (HCI). Existing approaches have achieved significant results; however, limitations persist, with the quantity and diversity of data being mo...

The Mason-Alberta Phonetic Segmenter: a forced alignment system based on deep neural networks and interpolation.

Phonetica
Given an orthographic transcription, forced alignment systems automatically determine boundaries between segments in speech, facilitating the use of large corpora. In the present paper, we introduce a neural network-based forced alignment system, the...