Dysarthria detection based on a deep learning model with a clinically-interpretable layer.

Journal: JASA express letters
Published Date:

Abstract

Studies have shown deep neural networks (DNN) as a potential tool for classifying dysarthric speakers and controls. However, representations used to train DNNs are largely not clinically interpretable, which limits clinical value. Here, a model with a bottleneck layer is trained to jointly learn a classification label and four clinically-interpretable features. Evaluation of two dysarthria subtypes shows that the proposed method can flexibly trade-off between improved classification accuracy and discovery of clinically-interpretable deficit patterns. The analysis using Shapley additive explanation shows the model learns a representation consistent with the disturbances that define the two dysarthria subtypes considered in this work.

Authors

  • Lingfeng Xu
    Department of Electronic Science and Technology, University of Science and Technology of China, Hefei 230027, China. xlfustc@mail.ustc.edu.cn.
  • Julie Liss
    College of Health Solutions, Arizona State University, Tempe, Arizona 85281, USA lingfen3@asu.edu; julie.liss@asu.edu; visar@asu.edu.
  • Visar Berisha