Cargando…

A study on computer vision for facial emotion recognition

Artificial intelligence has been successfully applied in various fields, one of which is computer vision. In this study, a deep neural network (DNN) was adopted for Facial emotion recognition (FER). One of the objectives in this study is to identify the critical facial features on which the DNN mode...

Descripción completa

Detalles Bibliográficos
Autores principales: Huang, Zi-Yu, Chiang, Chia-Chin, Chen, Jian-Hao, Chen, Yi-Chian, Chung, Hsin-Lung, Cai, Yu-Ping, Hsu, Hsiu-Chuan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10209161/
https://www.ncbi.nlm.nih.gov/pubmed/37225755
http://dx.doi.org/10.1038/s41598-023-35446-4
_version_ 1785046818156969984
author Huang, Zi-Yu
Chiang, Chia-Chin
Chen, Jian-Hao
Chen, Yi-Chian
Chung, Hsin-Lung
Cai, Yu-Ping
Hsu, Hsiu-Chuan
author_facet Huang, Zi-Yu
Chiang, Chia-Chin
Chen, Jian-Hao
Chen, Yi-Chian
Chung, Hsin-Lung
Cai, Yu-Ping
Hsu, Hsiu-Chuan
author_sort Huang, Zi-Yu
collection PubMed
description Artificial intelligence has been successfully applied in various fields, one of which is computer vision. In this study, a deep neural network (DNN) was adopted for Facial emotion recognition (FER). One of the objectives in this study is to identify the critical facial features on which the DNN model focuses for FER. In particular, we utilized a convolutional neural network (CNN), the combination of squeeze-and-excitation network and the residual neural network, for the task of FER. We utilized AffectNet and the Real-World Affective Faces Database (RAF-DB) as the facial expression databases that provide learning samples for the CNN. The feature maps were extracted from the residual blocks for further analysis. Our analysis shows that the features around the nose and mouth are critical facial landmarks for the neural networks. Cross-database validations were conducted between the databases. The network model trained on AffectNet achieved 77.37% accuracy when validated on the RAF-DB, while the network model pretrained on AffectNet and then transfer learned on the RAF-DB results in validation accuracy of 83.37%. The outcomes of this study would improve the understanding of neural networks and assist with improving computer vision accuracy.
format Online
Article
Text
id pubmed-10209161
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-102091612023-05-26 A study on computer vision for facial emotion recognition Huang, Zi-Yu Chiang, Chia-Chin Chen, Jian-Hao Chen, Yi-Chian Chung, Hsin-Lung Cai, Yu-Ping Hsu, Hsiu-Chuan Sci Rep Article Artificial intelligence has been successfully applied in various fields, one of which is computer vision. In this study, a deep neural network (DNN) was adopted for Facial emotion recognition (FER). One of the objectives in this study is to identify the critical facial features on which the DNN model focuses for FER. In particular, we utilized a convolutional neural network (CNN), the combination of squeeze-and-excitation network and the residual neural network, for the task of FER. We utilized AffectNet and the Real-World Affective Faces Database (RAF-DB) as the facial expression databases that provide learning samples for the CNN. The feature maps were extracted from the residual blocks for further analysis. Our analysis shows that the features around the nose and mouth are critical facial landmarks for the neural networks. Cross-database validations were conducted between the databases. The network model trained on AffectNet achieved 77.37% accuracy when validated on the RAF-DB, while the network model pretrained on AffectNet and then transfer learned on the RAF-DB results in validation accuracy of 83.37%. The outcomes of this study would improve the understanding of neural networks and assist with improving computer vision accuracy. Nature Publishing Group UK 2023-05-24 /pmc/articles/PMC10209161/ /pubmed/37225755 http://dx.doi.org/10.1038/s41598-023-35446-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Huang, Zi-Yu
Chiang, Chia-Chin
Chen, Jian-Hao
Chen, Yi-Chian
Chung, Hsin-Lung
Cai, Yu-Ping
Hsu, Hsiu-Chuan
A study on computer vision for facial emotion recognition
title A study on computer vision for facial emotion recognition
title_full A study on computer vision for facial emotion recognition
title_fullStr A study on computer vision for facial emotion recognition
title_full_unstemmed A study on computer vision for facial emotion recognition
title_short A study on computer vision for facial emotion recognition
title_sort study on computer vision for facial emotion recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10209161/
https://www.ncbi.nlm.nih.gov/pubmed/37225755
http://dx.doi.org/10.1038/s41598-023-35446-4
work_keys_str_mv AT huangziyu astudyoncomputervisionforfacialemotionrecognition
AT chiangchiachin astudyoncomputervisionforfacialemotionrecognition
AT chenjianhao astudyoncomputervisionforfacialemotionrecognition
AT chenyichian astudyoncomputervisionforfacialemotionrecognition
AT chunghsinlung astudyoncomputervisionforfacialemotionrecognition
AT caiyuping astudyoncomputervisionforfacialemotionrecognition
AT hsuhsiuchuan astudyoncomputervisionforfacialemotionrecognition
AT huangziyu studyoncomputervisionforfacialemotionrecognition
AT chiangchiachin studyoncomputervisionforfacialemotionrecognition
AT chenjianhao studyoncomputervisionforfacialemotionrecognition
AT chenyichian studyoncomputervisionforfacialemotionrecognition
AT chunghsinlung studyoncomputervisionforfacialemotionrecognition
AT caiyuping studyoncomputervisionforfacialemotionrecognition
AT hsuhsiuchuan studyoncomputervisionforfacialemotionrecognition