AIMC Topic: Speech

Clear Filters Showing 91 to 100 of 353 articles

Speech Emotion Recognition Using Convolution Neural Networks and Multi-Head Convolutional Transformer.

Sensors (Basel, Switzerland)
Speech emotion recognition (SER) is a challenging task in human-computer interaction (HCI) systems. One of the key challenges in speech emotion recognition is to extract the emotional features effectively from a speech utterance. Despite the promisin...

Efficient Self-Attention Model for Speech Recognition-Based Assistive Robots Control.

Sensors (Basel, Switzerland)
Assistive robots are tools that people living with upper body disabilities can leverage to autonomously perform Activities of Daily Living (ADL). Unfortunately, conventional control methods still rely on low-dimensional, easy-to-implement interfaces ...

A Speech Recognition Method Based on Domain-Specific Datasets and Confidence Decision Networks.

Sensors (Basel, Switzerland)
This paper proposes a speech recognition method based on a domain-specific language speech network (DSL-Net) and a confidence decision network (CD-Net). The method involves automatically training a domain-specific dataset, using pre-trained model par...

Block-level dependency syntax based model for end-to-end aspect-based sentiment analysis.

Neural networks : the official journal of the International Neural Network Society
End-to-End aspect-based sentiment analysis (E2E-ABSA) aims to jointly extract aspect terms and identify their sentiment polarities. Although previous research has demonstrated that syntax knowledge can be beneficial for E2E-ABSA, standard syntax depe...

Advancing Stuttering Detection via Data Augmentation, Class-Balanced Loss and Multi-Contextual Deep Learning.

IEEE journal of biomedical and health informatics
Stuttering is a neuro-developmental speech impairment characterized by uncontrolled utterances (interjections) and core behaviors (blocks, repetitions, and prolongations), and is caused by the failure of speech sensorimotors. Due to its complex natur...

Decoding Silent Speech Based on High-Density Surface Electromyogram Using Spatiotemporal Neural Network.

IEEE transactions on neural systems and rehabilitation engineering : a publication of the IEEE Engineering in Medicine and Biology Society
Finer-grained decoding at a phoneme or syllable level is a key technology for continuous recognition of silent speech based on surface electromyogram (sEMG). This paper aims at developing a novel syllable-level decoding method for continuous silent s...

Encoding of speech in convolutional layers and the brain stem based on language experience.

Scientific reports
Comparing artificial neural networks with outputs of neuroimaging techniques has recently seen substantial advances in (computer) vision and text-based language models. Here, we propose a framework to compare biological and artificial neural computat...

An octonion-based nonlinear echo state network for speech emotion recognition in Metaverse.

Neural networks : the official journal of the International Neural Network Society
While the Metaverse is becoming a popular trend and drawing much attention from academia, society, and businesses, processing cores used in its infrastructures need to be improved, particularly in terms of signal processing and pattern recognition. A...

Evidence of a predictive coding hierarchy in the human brain listening to speech.

Nature human behaviour
Considerable progress has recently been made in natural language processing: deep learning algorithms are increasingly able to generate, summarize, translate and classify texts. Yet, these language models still fail to match the language abilities of...

Robot leadership-Investigating human perceptions and reactions towards social robots showing leadership behaviors.

PloS one
Human-robot interaction research has shown that social robots can interact with humans in complex social situations and display leadership-related behaviors. Therefore, social robots could be able to take on leadership roles. The aim of our study was...