AIMC Topic: Speech

Clear Filters Showing 51 to 60 of 352 articles

Human-AI collaboration improves adults' oral biomechanical functions: A multi-centre, self-controlled clinical trial.

Journal of dentistry
OBJECTIVES: Maintenance of oral muscle functions is important for survival and communication. Utilizing Artificial Intelligence (AI) as a self-health-management material has shown promise. Here we developed a functional and AI-enabled smartphone e-Or...

Effective Phoneme Decoding With Hyperbolic Neural Networks for High-Performance Speech BCIs.

IEEE transactions on neural systems and rehabilitation engineering : a publication of the IEEE Engineering in Medicine and Biology Society
OBJECTIVE: Speech brain-computer interfaces (speech BCIs), which convert brain signals into spoken words or sentences, have demonstrated great potential for high-performance BCI communication. Phonemes are the basic pronunciation units. For monosylla...

Machine learning-based classification of Parkinson's disease using acoustic features: Insights from multilingual speech tasks.

Computers in biology and medicine
This study advances the automation of Parkinson's disease (PD) diagnosis by analyzing speech characteristics, leveraging a comprehensive approach that integrates a voting-based machine learning model. Given the growing prevalence of PD, especially am...

A Combined CNN Architecture for Speech Emotion Recognition.

Sensors (Basel, Switzerland)
Emotion recognition through speech is a technique employed in various scenarios of Human-Computer Interaction (HCI). Existing approaches have achieved significant results; however, limitations persist, with the quantity and diversity of data being mo...

The Mason-Alberta Phonetic Segmenter: a forced alignment system based on deep neural networks and interpolation.

Phonetica
Given an orthographic transcription, forced alignment systems automatically determine boundaries between segments in speech, facilitating the use of large corpora. In the present paper, we introduce a neural network-based forced alignment system, the...

Speech synthesis from three-axis accelerometer signals using conformer-based deep neural network.

Computers in biology and medicine
Silent speech interfaces (SSIs) have emerged as innovative non-acoustic communication methods, and our previous study demonstrated the significant potential of three-axis accelerometer-based SSIs to identify silently spoken words with high classifica...

A dual-region speech enhancement method based on voiceprint segmentation.

Neural networks : the official journal of the International Neural Network Society
Single-channel speech enhancement primarily relies on deep learning models to recover clean speech signals from noise-contaminated speech. These models establish a mapping relationship between noisy and clean speech. However, considering the sparse d...

Deep learning approach for dysphagia detection by syllable-based speech analysis with daily conversations.

Scientific reports
Dysphagia, a disorder affecting the ability to swallow, has a high prevalence among the older adults and can lead to serious health complications. Therefore, early detection of dysphagia is important. This study evaluated the effectiveness of a newly...

Multilevel hybrid handcrafted feature extraction based depression recognition method using speech.

Journal of affective disorders
BACKGROUND AND PURPOSE: Diagnosis of depression is based on tests performed by psychiatrists and information provided by patients or their relatives. In the field of machine learning (ML), numerous models have been devised to detect depression automa...

Deep Learning for Visual Speech Analysis: A Survey.

IEEE transactions on pattern analysis and machine intelligence
Visual speech, referring to the visual domain of speech, has attracted increasing attention due to its wide applications, such as public security, medical treatment, military defense, and film entertainment. As a powerful AI strategy, deep learning t...