Cargando…

Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model

Autism spectrum disorder (ASD) is diagnosed on the basis of speech and communication differences, amongst other symptoms. Since conversations are essential for building connections with others, it is important to understand the exact nature of differences between autistic and non-autistic verbal beh...

Descripción completa

Detalles Bibliográficos
Autores principales: Plank, I. S., Koehler, J. C., Nelson, A. M., Koutsouleris, N., Falter-Wagner, C. M.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10658003/
https://www.ncbi.nlm.nih.gov/pubmed/38025455
http://dx.doi.org/10.3389/fpsyt.2023.1257569
_version_ 1785137339804155904
author Plank, I. S.
Koehler, J. C.
Nelson, A. M.
Koutsouleris, N.
Falter-Wagner, C. M.
author_facet Plank, I. S.
Koehler, J. C.
Nelson, A. M.
Koutsouleris, N.
Falter-Wagner, C. M.
author_sort Plank, I. S.
collection PubMed
description Autism spectrum disorder (ASD) is diagnosed on the basis of speech and communication differences, amongst other symptoms. Since conversations are essential for building connections with others, it is important to understand the exact nature of differences between autistic and non-autistic verbal behaviour and evaluate the potential of these differences for diagnostics. In this study, we recorded dyadic conversations and used automated extraction of speech and interactional turn-taking features of 54 non-autistic and 26 autistic participants. The extracted speech and turn-taking parameters showed high potential as a diagnostic marker. A linear support vector machine was able to predict the dyad type with 76.2% balanced accuracy (sensitivity: 73.8%, specificity: 78.6%), suggesting that digitally assisted diagnostics could significantly enhance the current clinical diagnostic process due to their objectivity and scalability. In group comparisons on the individual and dyadic level, we found that autistic interaction partners talked slower and in a more monotonous manner than non-autistic interaction partners and that mixed dyads consisting of an autistic and a non-autistic participant had increased periods of silence, and the intensity, i.e. loudness, of their speech was more synchronous.
format Online
Article
Text
id pubmed-10658003
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-106580032023-11-06 Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model Plank, I. S. Koehler, J. C. Nelson, A. M. Koutsouleris, N. Falter-Wagner, C. M. Front Psychiatry Psychiatry Autism spectrum disorder (ASD) is diagnosed on the basis of speech and communication differences, amongst other symptoms. Since conversations are essential for building connections with others, it is important to understand the exact nature of differences between autistic and non-autistic verbal behaviour and evaluate the potential of these differences for diagnostics. In this study, we recorded dyadic conversations and used automated extraction of speech and interactional turn-taking features of 54 non-autistic and 26 autistic participants. The extracted speech and turn-taking parameters showed high potential as a diagnostic marker. A linear support vector machine was able to predict the dyad type with 76.2% balanced accuracy (sensitivity: 73.8%, specificity: 78.6%), suggesting that digitally assisted diagnostics could significantly enhance the current clinical diagnostic process due to their objectivity and scalability. In group comparisons on the individual and dyadic level, we found that autistic interaction partners talked slower and in a more monotonous manner than non-autistic interaction partners and that mixed dyads consisting of an autistic and a non-autistic participant had increased periods of silence, and the intensity, i.e. loudness, of their speech was more synchronous. Frontiers Media S.A. 2023-11-06 /pmc/articles/PMC10658003/ /pubmed/38025455 http://dx.doi.org/10.3389/fpsyt.2023.1257569 Text en Copyright © 2023 Plank, Koehler, Nelson, Koutsouleris and Falter-Wagner. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychiatry
Plank, I. S.
Koehler, J. C.
Nelson, A. M.
Koutsouleris, N.
Falter-Wagner, C. M.
Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
title Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
title_full Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
title_fullStr Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
title_full_unstemmed Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
title_short Automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
title_sort automated extraction of speech and turn-taking parameters in autism allows for diagnostic classification using a multivariable prediction model
topic Psychiatry
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10658003/
https://www.ncbi.nlm.nih.gov/pubmed/38025455
http://dx.doi.org/10.3389/fpsyt.2023.1257569
work_keys_str_mv AT plankis automatedextractionofspeechandturntakingparametersinautismallowsfordiagnosticclassificationusingamultivariablepredictionmodel
AT koehlerjc automatedextractionofspeechandturntakingparametersinautismallowsfordiagnosticclassificationusingamultivariablepredictionmodel
AT nelsonam automatedextractionofspeechandturntakingparametersinautismallowsfordiagnosticclassificationusingamultivariablepredictionmodel
AT koutsoulerisn automatedextractionofspeechandturntakingparametersinautismallowsfordiagnosticclassificationusingamultivariablepredictionmodel
AT falterwagnercm automatedextractionofspeechandturntakingparametersinautismallowsfordiagnosticclassificationusingamultivariablepredictionmodel