Cargando…

Multimodal explainable AI predicts upcoming speech behavior in adults who stutter

A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body...

Descripción completa

Detalles Bibliográficos
Autores principales: Das, Arun, Mock, Jeffrey, Irani, Farzan, Huang, Yufei, Najafirad, Peyman, Golob, Edward
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9376608/
https://www.ncbi.nlm.nih.gov/pubmed/35979337
http://dx.doi.org/10.3389/fnins.2022.912798
_version_ 1784768182964191232
author Das, Arun
Mock, Jeffrey
Irani, Farzan
Huang, Yufei
Najafirad, Peyman
Golob, Edward
author_facet Das, Arun
Mock, Jeffrey
Irani, Farzan
Huang, Yufei
Najafirad, Peyman
Golob, Edward
author_sort Das, Arun
collection PubMed
description A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body. Autonomic nervous system activity is the best-known example, but, muscles in the eyes and face can also index brain activity. Mostly parallel lines of artificial intelligence research show that EEG and facial muscles both encode information about emotion, pain, attention, and social interactions, among other topics. In this study, we examined adults who stutter (AWS) to understand the relations between dynamic brain and facial muscle activity and predictions about future behavior (fluent or stuttered speech). AWS can provide insight into brain-behavior dynamics because they naturally fluctuate between episodes of fluent and stuttered speech behavior. We focused on the period when speech preparation occurs, and used EEG and facial muscle activity measured from video to predict whether the upcoming speech would be fluent or stuttered. An explainable self-supervised multimodal architecture learned the temporal dynamics of both EEG and facial muscle movements during speech preparation in AWS, and predicted fluent or stuttered speech at 80.8% accuracy (chance=50%). Specific EEG and facial muscle signals distinguished fluent and stuttered trials, and systematically varied from early to late speech preparation time periods. The self-supervised architecture successfully identified multimodal activity that predicted upcoming behavior on a trial-by-trial basis. This approach could be applied to understanding the neural mechanisms driving variable behavior and symptoms in a wide range of neurological and psychiatric disorders. The combination of direct measures of neural activity and simple video data may be applied to developing technologies that estimate brain state from subtle bodily signals.
format Online
Article
Text
id pubmed-9376608
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-93766082022-08-16 Multimodal explainable AI predicts upcoming speech behavior in adults who stutter Das, Arun Mock, Jeffrey Irani, Farzan Huang, Yufei Najafirad, Peyman Golob, Edward Front Neurosci Neuroscience A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body. Autonomic nervous system activity is the best-known example, but, muscles in the eyes and face can also index brain activity. Mostly parallel lines of artificial intelligence research show that EEG and facial muscles both encode information about emotion, pain, attention, and social interactions, among other topics. In this study, we examined adults who stutter (AWS) to understand the relations between dynamic brain and facial muscle activity and predictions about future behavior (fluent or stuttered speech). AWS can provide insight into brain-behavior dynamics because they naturally fluctuate between episodes of fluent and stuttered speech behavior. We focused on the period when speech preparation occurs, and used EEG and facial muscle activity measured from video to predict whether the upcoming speech would be fluent or stuttered. An explainable self-supervised multimodal architecture learned the temporal dynamics of both EEG and facial muscle movements during speech preparation in AWS, and predicted fluent or stuttered speech at 80.8% accuracy (chance=50%). Specific EEG and facial muscle signals distinguished fluent and stuttered trials, and systematically varied from early to late speech preparation time periods. The self-supervised architecture successfully identified multimodal activity that predicted upcoming behavior on a trial-by-trial basis. This approach could be applied to understanding the neural mechanisms driving variable behavior and symptoms in a wide range of neurological and psychiatric disorders. The combination of direct measures of neural activity and simple video data may be applied to developing technologies that estimate brain state from subtle bodily signals. Frontiers Media S.A. 2022-08-01 /pmc/articles/PMC9376608/ /pubmed/35979337 http://dx.doi.org/10.3389/fnins.2022.912798 Text en Copyright © 2022 Das, Mock, Irani, Huang, Najafirad and Golob. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Das, Arun
Mock, Jeffrey
Irani, Farzan
Huang, Yufei
Najafirad, Peyman
Golob, Edward
Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
title Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
title_full Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
title_fullStr Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
title_full_unstemmed Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
title_short Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
title_sort multimodal explainable ai predicts upcoming speech behavior in adults who stutter
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9376608/
https://www.ncbi.nlm.nih.gov/pubmed/35979337
http://dx.doi.org/10.3389/fnins.2022.912798
work_keys_str_mv AT dasarun multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter
AT mockjeffrey multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter
AT iranifarzan multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter
AT huangyufei multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter
AT najafiradpeyman multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter
AT golobedward multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter