Cargando…
Dysarthria detection based on a deep learning model with a clinically-interpretable layer
Studies have shown deep neural networks (DNN) as a potential tool for classifying dysarthric speakers and controls. However, representations used to train DNNs are largely not clinically interpretable, which limits clinical value. Here, a model with a bottleneck layer is trained to jointly learn a c...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Acoustical Society of America
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9835557/ https://www.ncbi.nlm.nih.gov/pubmed/36725533 http://dx.doi.org/10.1121/10.0016833 |
_version_ | 1784868692554678272 |
---|---|
author | Xu, Lingfeng Liss, Julie Berisha, Visar |
author_facet | Xu, Lingfeng Liss, Julie Berisha, Visar |
author_sort | Xu, Lingfeng |
collection | PubMed |
description | Studies have shown deep neural networks (DNN) as a potential tool for classifying dysarthric speakers and controls. However, representations used to train DNNs are largely not clinically interpretable, which limits clinical value. Here, a model with a bottleneck layer is trained to jointly learn a classification label and four clinically-interpretable features. Evaluation of two dysarthria subtypes shows that the proposed method can flexibly trade-off between improved classification accuracy and discovery of clinically-interpretable deficit patterns. The analysis using Shapley additive explanation shows the model learns a representation consistent with the disturbances that define the two dysarthria subtypes considered in this work. |
format | Online Article Text |
id | pubmed-9835557 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Acoustical Society of America |
record_format | MEDLINE/PubMed |
spelling | pubmed-98355572023-01-13 Dysarthria detection based on a deep learning model with a clinically-interpretable layer Xu, Lingfeng Liss, Julie Berisha, Visar JASA Express Lett Speech Communication Studies have shown deep neural networks (DNN) as a potential tool for classifying dysarthric speakers and controls. However, representations used to train DNNs are largely not clinically interpretable, which limits clinical value. Here, a model with a bottleneck layer is trained to jointly learn a classification label and four clinically-interpretable features. Evaluation of two dysarthria subtypes shows that the proposed method can flexibly trade-off between improved classification accuracy and discovery of clinically-interpretable deficit patterns. The analysis using Shapley additive explanation shows the model learns a representation consistent with the disturbances that define the two dysarthria subtypes considered in this work. Acoustical Society of America 2023-01 2023-01-10 /pmc/articles/PMC9835557/ /pubmed/36725533 http://dx.doi.org/10.1121/10.0016833 Text en © 2023 Author(s). 2691-1191/2021/3(1)/015201/8 https://creativecommons.org/licenses/by/4.0/All article content, except where otherwise noted, is licensed under a Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ). |
spellingShingle | Speech Communication Xu, Lingfeng Liss, Julie Berisha, Visar Dysarthria detection based on a deep learning model with a clinically-interpretable layer |
title | Dysarthria detection based on a deep learning model with a clinically-interpretable layer |
title_full | Dysarthria detection based on a deep learning model with a clinically-interpretable layer |
title_fullStr | Dysarthria detection based on a deep learning model with a clinically-interpretable layer |
title_full_unstemmed | Dysarthria detection based on a deep learning model with a clinically-interpretable layer |
title_short | Dysarthria detection based on a deep learning model with a clinically-interpretable layer |
title_sort | dysarthria detection based on a deep learning model with a clinically-interpretable layer |
topic | Speech Communication |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9835557/ https://www.ncbi.nlm.nih.gov/pubmed/36725533 http://dx.doi.org/10.1121/10.0016833 |
work_keys_str_mv | AT xulingfeng dysarthriadetectionbasedonadeeplearningmodelwithaclinicallyinterpretablelayer AT lissjulie dysarthriadetectionbasedonadeeplearningmodelwithaclinicallyinterpretablelayer AT berishavisar dysarthriadetectionbasedonadeeplearningmodelwithaclinicallyinterpretablelayer |