Cargando…
Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study
BACKGROUND: Implementing automated facial expression recognition on mobile devices could provide an accessible diagnostic and therapeutic tool for those who struggle to recognize facial expressions, including children with developmental behavioral conditions such as autism. Despite recent advances i...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
JMIR Publications
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10131663/ https://www.ncbi.nlm.nih.gov/pubmed/35962462 http://dx.doi.org/10.2196/39917 |
_version_ | 1785031224530567168 |
---|---|
author | Banerjee, Agnik Mutlu, Onur Cezmi Kline, Aaron Surabhi, Saimourya Washington, Peter Wall, Dennis Paul |
author_facet | Banerjee, Agnik Mutlu, Onur Cezmi Kline, Aaron Surabhi, Saimourya Washington, Peter Wall, Dennis Paul |
author_sort | Banerjee, Agnik |
collection | PubMed |
description | BACKGROUND: Implementing automated facial expression recognition on mobile devices could provide an accessible diagnostic and therapeutic tool for those who struggle to recognize facial expressions, including children with developmental behavioral conditions such as autism. Despite recent advances in facial expression classifiers for children, existing models are too computationally expensive for smartphone use. OBJECTIVE: We explored several state-of-the-art facial expression classifiers designed for mobile devices, used posttraining optimization techniques for both classification performance and efficiency on a Motorola Moto G6 phone, evaluated the importance of training our classifiers on children versus adults, and evaluated the models’ performance against different ethnic groups. METHODS: We collected images from 12 public data sets and used video frames crowdsourced from the GuessWhat app to train our classifiers. All images were annotated for 7 expressions: neutral, fear, happiness, sadness, surprise, anger, and disgust. We tested 3 copies for each of 5 different convolutional neural network architectures: MobileNetV3-Small 1.0x, MobileNetV2 1.0x, EfficientNetB0, MobileNetV3-Large 1.0x, and NASNetMobile. We trained the first copy on images of children, second copy on images of adults, and third copy on all data sets. We evaluated each model against the entire Child Affective Facial Expression (CAFE) set and by ethnicity. We performed weight pruning, weight clustering, and quantize-aware training when possible and profiled each model’s performance on the Moto G6. RESULTS: Our best model, a MobileNetV3-Large network pretrained on ImageNet, achieved 65.78% accuracy and 65.31% F(1)-score on the CAFE and a 90-millisecond inference latency on a Moto G6 phone when trained on all data. This accuracy is only 1.12% lower than the current state of the art for CAFE, a model with 13.91x more parameters that was unable to run on the Moto G6 due to its size, even when fully optimized. When trained solely on children, this model achieved 60.57% accuracy and 60.29% F(1)-score. When trained only on adults, the model received 53.36% accuracy and 53.10% F(1)-score. Although the MobileNetV3-Large trained on all data sets achieved nearly a 60% F(1)-score across all ethnicities, the data sets for South Asian and African American children achieved lower accuracy (as much as 11.56%) and F(1)-score (as much as 11.25%) than other groups. CONCLUSIONS: With specialized design and optimization techniques, facial expression classifiers can become lightweight enough to run on mobile devices and achieve state-of-the-art performance. There is potentially a “data shift” phenomenon between facial expressions of children compared with adults; our classifiers performed much better when trained on children. Certain underrepresented ethnic groups (e.g., South Asian and African American) also perform significantly worse than groups such as European Caucasian despite similar data quality. Our models can be integrated into mobile health therapies to help diagnose autism spectrum disorder and provide targeted therapeutic treatment to children. |
format | Online Article Text |
id | pubmed-10131663 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | JMIR Publications |
record_format | MEDLINE/PubMed |
spelling | pubmed-101316632023-04-27 Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study Banerjee, Agnik Mutlu, Onur Cezmi Kline, Aaron Surabhi, Saimourya Washington, Peter Wall, Dennis Paul JMIR Form Res Original Paper BACKGROUND: Implementing automated facial expression recognition on mobile devices could provide an accessible diagnostic and therapeutic tool for those who struggle to recognize facial expressions, including children with developmental behavioral conditions such as autism. Despite recent advances in facial expression classifiers for children, existing models are too computationally expensive for smartphone use. OBJECTIVE: We explored several state-of-the-art facial expression classifiers designed for mobile devices, used posttraining optimization techniques for both classification performance and efficiency on a Motorola Moto G6 phone, evaluated the importance of training our classifiers on children versus adults, and evaluated the models’ performance against different ethnic groups. METHODS: We collected images from 12 public data sets and used video frames crowdsourced from the GuessWhat app to train our classifiers. All images were annotated for 7 expressions: neutral, fear, happiness, sadness, surprise, anger, and disgust. We tested 3 copies for each of 5 different convolutional neural network architectures: MobileNetV3-Small 1.0x, MobileNetV2 1.0x, EfficientNetB0, MobileNetV3-Large 1.0x, and NASNetMobile. We trained the first copy on images of children, second copy on images of adults, and third copy on all data sets. We evaluated each model against the entire Child Affective Facial Expression (CAFE) set and by ethnicity. We performed weight pruning, weight clustering, and quantize-aware training when possible and profiled each model’s performance on the Moto G6. RESULTS: Our best model, a MobileNetV3-Large network pretrained on ImageNet, achieved 65.78% accuracy and 65.31% F(1)-score on the CAFE and a 90-millisecond inference latency on a Moto G6 phone when trained on all data. This accuracy is only 1.12% lower than the current state of the art for CAFE, a model with 13.91x more parameters that was unable to run on the Moto G6 due to its size, even when fully optimized. When trained solely on children, this model achieved 60.57% accuracy and 60.29% F(1)-score. When trained only on adults, the model received 53.36% accuracy and 53.10% F(1)-score. Although the MobileNetV3-Large trained on all data sets achieved nearly a 60% F(1)-score across all ethnicities, the data sets for South Asian and African American children achieved lower accuracy (as much as 11.56%) and F(1)-score (as much as 11.25%) than other groups. CONCLUSIONS: With specialized design and optimization techniques, facial expression classifiers can become lightweight enough to run on mobile devices and achieve state-of-the-art performance. There is potentially a “data shift” phenomenon between facial expressions of children compared with adults; our classifiers performed much better when trained on children. Certain underrepresented ethnic groups (e.g., South Asian and African American) also perform significantly worse than groups such as European Caucasian despite similar data quality. Our models can be integrated into mobile health therapies to help diagnose autism spectrum disorder and provide targeted therapeutic treatment to children. JMIR Publications 2023-03-21 /pmc/articles/PMC10131663/ /pubmed/35962462 http://dx.doi.org/10.2196/39917 Text en ©Agnik Banerjee, Onur Cezmi Mutlu, Aaron Kline, Saimourya Surabhi, Peter Washington, Dennis Paul Wall. Originally published in JMIR Formative Research (https://formative.jmir.org), 21.03.2023. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included. |
spellingShingle | Original Paper Banerjee, Agnik Mutlu, Onur Cezmi Kline, Aaron Surabhi, Saimourya Washington, Peter Wall, Dennis Paul Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study |
title | Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study |
title_full | Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study |
title_fullStr | Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study |
title_full_unstemmed | Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study |
title_short | Training and Profiling a Pediatric Facial Expression Classifier for Children on Mobile Devices: Machine Learning Study |
title_sort | training and profiling a pediatric facial expression classifier for children on mobile devices: machine learning study |
topic | Original Paper |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10131663/ https://www.ncbi.nlm.nih.gov/pubmed/35962462 http://dx.doi.org/10.2196/39917 |
work_keys_str_mv | AT banerjeeagnik trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy AT mutluonurcezmi trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy AT klineaaron trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy AT surabhisaimourya trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy AT washingtonpeter trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy AT walldennispaul trainingandprofilingapediatricfacialexpressionclassifierforchildrenonmobiledevicesmachinelearningstudy |