Cargando…

Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study

BACKGROUND: Implementing automated facial expression recognition on mobile devices could provide an accessible diagnostic and therapeutic tool for those who struggle to recognize facial expressions, including children with developmental behavioral conditions such as autism. Despite recent advances i...

Descripción completa

Detalles Bibliográficos
Autores principales: Banerjee, Agnik, Mutlu, Onur Cezmi, Kline, Aaron, Surabhi, Saimourya, Washington, Peter, Wall, Dennis Paul
Formato: Online Artículo Texto
Lenguaje:English
Publicado: JMIR Publications 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10131663/
https://www.ncbi.nlm.nih.gov/pubmed/35962462
http://dx.doi.org/10.2196/39917
_version_ 1785031224530567168
author Banerjee, Agnik
Mutlu, Onur Cezmi
Kline, Aaron
Surabhi, Saimourya
Washington, Peter
Wall, Dennis Paul
author_facet Banerjee, Agnik
Mutlu, Onur Cezmi
Kline, Aaron
Surabhi, Saimourya
Washington, Peter
Wall, Dennis Paul
author_sort Banerjee, Agnik
collection PubMed
description BACKGROUND: Implementing automated facial expression recognition on mobile devices could provide an accessible diagnostic and therapeutic tool for those who struggle to recognize facial expressions, including children with developmental behavioral conditions such as autism. Despite recent advances in facial expression classifiers for children, existing models are too computationally expensive for smartphone use. OBJECTIVE: We explored several state-of-the-art facial expression classifiers designed for mobile devices, used posttraining optimization techniques for both classification performance and efficiency on a Motorola Moto G6 phone, evaluated the importance of training our classifiers on children versus adults, and evaluated the models’ performance against different ethnic groups. METHODS: We collected images from 12 public data sets and used video frames crowdsourced from the GuessWhat app to train our classifiers. All images were annotated for 7 expressions: neutral, fear, happiness, sadness, surprise, anger, and disgust. We tested 3 copies for each of 5 different convolutional neural network architectures: MobileNetV3-Small 1.0x, MobileNetV2 1.0x, EfficientNetB0, MobileNetV3-Large 1.0x, and NASNetMobile. We trained the first copy on images of children, second copy on images of adults, and third copy on all data sets. We evaluated each model against the entire Child Affective Facial Expression (CAFE) set and by ethnicity. We performed weight pruning, weight clustering, and quantize-aware training when possible and profiled each model’s performance on the Moto G6. RESULTS: Our best model, a MobileNetV3-Large network pretrained on ImageNet, achieved 65.78% accuracy and 65.31% F(1)-score on the CAFE and a 90-millisecond inference latency on a Moto G6 phone when trained on all data. This accuracy is only 1.12% lower than the current state of the art for CAFE, a model with 13.91x more parameters that was unable to run on the Moto G6 due to its size, even when fully optimized. When trained solely on children, this model achieved 60.57% accuracy and 60.29% F(1)-score. When trained only on adults, the model received 53.36% accuracy and 53.10% F(1)-score. Although the MobileNetV3-Large trained on all data sets achieved nearly a 60% F(1)-score across all ethnicities, the data sets for South Asian and African American children achieved lower accuracy (as much as 11.56%) and F(1)-score (as much as 11.25%) than other groups. CONCLUSIONS: With specialized design and optimization techniques, facial expression classifiers can become lightweight enough to run on mobile devices and achieve state-of-the-art performance. There is potentially a “data shift” phenomenon between facial expressions of children compared with adults; our classifiers performed much better when trained on children. Certain underrepresented ethnic groups (e.g., South Asian and African American) also perform significantly worse than groups such as European Caucasian despite similar data quality. Our models can be integrated into mobile health therapies to help diagnose autism spectrum disorder and provide targeted therapeutic treatment to children.
format Online
Article
Text
id pubmed-10131663
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher JMIR Publications
record_format MEDLINE/PubMed
spelling pubmed-101316632023-04-27 Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study Banerjee, Agnik Mutlu, Onur Cezmi Kline, Aaron Surabhi, Saimourya Washington, Peter Wall, Dennis Paul JMIR Form Res Original Paper BACKGROUND: Implementing automated facial expression recognition on mobile devices could provide an accessible diagnostic and therapeutic tool for those who struggle to recognize facial expressions, including children with developmental behavioral conditions such as autism. Despite recent advances in facial expression classifiers for children, existing models are too computationally expensive for smartphone use. OBJECTIVE: We explored several state-of-the-art facial expression classifiers designed for mobile devices, used posttraining optimization techniques for both classification performance and efficiency on a Motorola Moto G6 phone, evaluated the importance of training our classifiers on children versus adults, and evaluated the models’ performance against different ethnic groups. METHODS: We collected images from 12 public data sets and used video frames crowdsourced from the GuessWhat app to train our classifiers. All images were annotated for 7 expressions: neutral, fear, happiness, sadness, surprise, anger, and disgust. We tested 3 copies for each of 5 different convolutional neural network architectures: MobileNetV3-Small 1.0x, MobileNetV2 1.0x, EfficientNetB0, MobileNetV3-Large 1.0x, and NASNetMobile. We trained the first copy on images of children, second copy on images of adults, and third copy on all data sets. We evaluated each model against the entire Child Affective Facial Expression (CAFE) set and by ethnicity. We performed weight pruning, weight clustering, and quantize-aware training when possible and profiled each model’s performance on the Moto G6. RESULTS: Our best model, a MobileNetV3-Large network pretrained on ImageNet, achieved 65.78% accuracy and 65.31% F(1)-score on the CAFE and a 90-millisecond inference latency on a Moto G6 phone when trained on all data. This accuracy is only 1.12% lower than the current state of the art for CAFE, a model with 13.91x more parameters that was unable to run on the Moto G6 due to its size, even when fully optimized. When trained solely on children, this model achieved 60.57% accuracy and 60.29% F(1)-score. When trained only on adults, the model received 53.36% accuracy and 53.10% F(1)-score. Although the MobileNetV3-Large trained on all data sets achieved nearly a 60% F(1)-score across all ethnicities, the data sets for South Asian and African American children achieved lower accuracy (as much as 11.56%) and F(1)-score (as much as 11.25%) than other groups. CONCLUSIONS: With specialized design and optimization techniques, facial expression classifiers can become lightweight enough to run on mobile devices and achieve state-of-the-art performance. There is potentially a “data shift” phenomenon between facial expressions of children compared with adults; our classifiers performed much better when trained on children. Certain underrepresented ethnic groups (e.g., South Asian and African American) also perform significantly worse than groups such as European Caucasian despite similar data quality. Our models can be integrated into mobile health therapies to help diagnose autism spectrum disorder and provide targeted therapeutic treatment to children. JMIR Publications 2023-03-21 /pmc/articles/PMC10131663/ /pubmed/35962462 http://dx.doi.org/10.2196/39917 Text en ©Agnik Banerjee, Onur Cezmi Mutlu, Aaron Kline, Saimourya Surabhi, Peter Washington, Dennis Paul Wall. Originally published in JMIR Formative Research (https://formative.jmir.org), 21.03.2023. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.
spellingShingle Original Paper
Banerjee, Agnik
Mutlu, Onur Cezmi
Kline, Aaron
Surabhi, Saimourya
Washington, Peter
Wall, Dennis Paul
Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
title Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
title_full Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
title_fullStr Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
title_full_unstemmed Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
title_short Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
title_sort training and profiling a pediatric facial expression classifier for children on mobile devices: machine learning study
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10131663/
https://www.ncbi.nlm.nih.gov/pubmed/35962462
http://dx.doi.org/10.2196/39917
work_keys_str_mv AT banerjeeagnik trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy
AT mutluonurcezmi trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy
AT klineaaron trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy
AT surabhisaimourya trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy
AT washingtonpeter trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy
AT walldennispaul trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy