AIMC Topic: Speech Production Measurement

Clear Filters Showing 1 to 10 of 40 articles

Computing nasalance with MFCCs and Convolutional Neural Networks.

PloS one
Nasalance is a valuable clinical biomarker for hypernasality. It is computed as the ratio of acoustic energy emitted through the nose to the total energy emitted through the mouth and nose (eNasalance). A new approach is proposed to compute nasalance...

Artificial Intelligence-Assisted Speech Therapy for /ɹ/: A Single-Case Experimental Study.

American journal of speech-language pathology
PURPOSE: This feasibility trial describes changes in rhotic production in residual speech sound disorder following ten 40-min sessions including artificial intelligence (AI)-assisted motor-based intervention with ChainingAI, a version of Speech Motor...

Accuracy of Speech Sound Analysis: Comparison of an Automatic Artificial Intelligence Algorithm With Clinician Assessment.

Journal of speech, language, and hearing research : JSLHR
PURPOSE: Automatic speech analysis (ASA) and automatic speech recognition systems are increasingly being used in the treatment of speech sound disorders (SSDs). When utilized as a home practice tool or in the absence of the clinician, the ASA system ...

TranStutter: A Convolution-Free Transformer-Based Deep Learning Method to Classify Stuttered Speech Using 2D Mel-Spectrogram Visualization and Attention-Based Feature Representation.

Sensors (Basel, Switzerland)
Stuttering, a prevalent neurodevelopmental disorder, profoundly affects fluent speech, causing involuntary interruptions and recurrent sound patterns. This study addresses the critical need for the accurate classification of stuttering types. The res...

Recognition of the Effect of Vocal Exercises by Fuzzy Triangular Naive Bayes, a Machine Learning Classifier: A Preliminary Analysis.

Journal of voice : official journal of the Voice Foundation
OBJECTIVES: Machine learning (ML) methods allow the development of expert systems for pattern recognition and predictive analysis of intervention outcomes. It has been used in Voice Sciences, mainly to discriminate between healthy and dysphonic voice...

Pathological Voice Detection Based on Phase Reconstitution and Convolutional Neural Network.

Journal of voice : official journal of the Voice Foundation
The nonlinear dynamic features can effectively describe the acoustic characteristics of normal and pathological voice. In this paper, the phase space reconstruction and convolution neural network are used to classify the normal and pathological voice...

Deep-Learning-Based Representation of Vocal Fold Dynamics in Adductor Spasmodic Dysphonia during Connected Speech in High-Speed Videoendoscopy.

Journal of voice : official journal of the Voice Foundation
OBJECTIVE: Adductor spasmodic dysphonia (AdSD) is a neurogenic dystonia, which causes spasms of the laryngeal muscles. This disorder mainly affects production of connected speech. To understand how AdSD affects vocal fold (VF) movements and hence, th...

Deep learning in automatic detection of dysphonia: Comparing acoustic features and developing a generalizable framework.

International journal of language & communication disorders
BACKGROUND: Auditory-perceptual assessment of voice is a subjective procedure. Artificial intelligence with deep learning (DL) may improve the consistency and accessibility of this task. It is unclear how a DL model performs on different acoustic fea...