Cargando…
Multimodal explainable AI predicts upcoming speech behavior in adults who stutter
A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9376608/ https://www.ncbi.nlm.nih.gov/pubmed/35979337 http://dx.doi.org/10.3389/fnins.2022.912798 |
_version_ | 1784768182964191232 |
---|---|
author | Das, Arun Mock, Jeffrey Irani, Farzan Huang, Yufei Najafirad, Peyman Golob, Edward |
author_facet | Das, Arun Mock, Jeffrey Irani, Farzan Huang, Yufei Najafirad, Peyman Golob, Edward |
author_sort | Das, Arun |
collection | PubMed |
description | A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body. Autonomic nervous system activity is the best-known example, but, muscles in the eyes and face can also index brain activity. Mostly parallel lines of artificial intelligence research show that EEG and facial muscles both encode information about emotion, pain, attention, and social interactions, among other topics. In this study, we examined adults who stutter (AWS) to understand the relations between dynamic brain and facial muscle activity and predictions about future behavior (fluent or stuttered speech). AWS can provide insight into brain-behavior dynamics because they naturally fluctuate between episodes of fluent and stuttered speech behavior. We focused on the period when speech preparation occurs, and used EEG and facial muscle activity measured from video to predict whether the upcoming speech would be fluent or stuttered. An explainable self-supervised multimodal architecture learned the temporal dynamics of both EEG and facial muscle movements during speech preparation in AWS, and predicted fluent or stuttered speech at 80.8% accuracy (chance=50%). Specific EEG and facial muscle signals distinguished fluent and stuttered trials, and systematically varied from early to late speech preparation time periods. The self-supervised architecture successfully identified multimodal activity that predicted upcoming behavior on a trial-by-trial basis. This approach could be applied to understanding the neural mechanisms driving variable behavior and symptoms in a wide range of neurological and psychiatric disorders. The combination of direct measures of neural activity and simple video data may be applied to developing technologies that estimate brain state from subtle bodily signals. |
format | Online Article Text |
id | pubmed-9376608 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-93766082022-08-16 Multimodal explainable AI predicts upcoming speech behavior in adults who stutter Das, Arun Mock, Jeffrey Irani, Farzan Huang, Yufei Najafirad, Peyman Golob, Edward Front Neurosci Neuroscience A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body. Autonomic nervous system activity is the best-known example, but, muscles in the eyes and face can also index brain activity. Mostly parallel lines of artificial intelligence research show that EEG and facial muscles both encode information about emotion, pain, attention, and social interactions, among other topics. In this study, we examined adults who stutter (AWS) to understand the relations between dynamic brain and facial muscle activity and predictions about future behavior (fluent or stuttered speech). AWS can provide insight into brain-behavior dynamics because they naturally fluctuate between episodes of fluent and stuttered speech behavior. We focused on the period when speech preparation occurs, and used EEG and facial muscle activity measured from video to predict whether the upcoming speech would be fluent or stuttered. An explainable self-supervised multimodal architecture learned the temporal dynamics of both EEG and facial muscle movements during speech preparation in AWS, and predicted fluent or stuttered speech at 80.8% accuracy (chance=50%). Specific EEG and facial muscle signals distinguished fluent and stuttered trials, and systematically varied from early to late speech preparation time periods. The self-supervised architecture successfully identified multimodal activity that predicted upcoming behavior on a trial-by-trial basis. This approach could be applied to understanding the neural mechanisms driving variable behavior and symptoms in a wide range of neurological and psychiatric disorders. The combination of direct measures of neural activity and simple video data may be applied to developing technologies that estimate brain state from subtle bodily signals. Frontiers Media S.A. 2022-08-01 /pmc/articles/PMC9376608/ /pubmed/35979337 http://dx.doi.org/10.3389/fnins.2022.912798 Text en Copyright © 2022 Das, Mock, Irani, Huang, Najafirad and Golob. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Das, Arun Mock, Jeffrey Irani, Farzan Huang, Yufei Najafirad, Peyman Golob, Edward Multimodal explainable AI predicts upcoming speech behavior in adults who stutter |
title | Multimodal explainable AI predicts upcoming speech behavior in adults who stutter |
title_full | Multimodal explainable AI predicts upcoming speech behavior in adults who stutter |
title_fullStr | Multimodal explainable AI predicts upcoming speech behavior in adults who stutter |
title_full_unstemmed | Multimodal explainable AI predicts upcoming speech behavior in adults who stutter |
title_short | Multimodal explainable AI predicts upcoming speech behavior in adults who stutter |
title_sort | multimodal explainable ai predicts upcoming speech behavior in adults who stutter |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9376608/ https://www.ncbi.nlm.nih.gov/pubmed/35979337 http://dx.doi.org/10.3389/fnins.2022.912798 |
work_keys_str_mv | AT dasarun multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter AT mockjeffrey multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter AT iranifarzan multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter AT huangyufei multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter AT najafiradpeyman multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter AT golobedward multimodalexplainableaipredictsupcomingspeechbehaviorinadultswhostutter |