Cargando…
A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data
Facial Expression Recognition (FER) can be widely applied to various research areas, such as mental diseases diagnosis and human social/physiological interaction detection. With the emerging advanced technologies in hardware and sensors, FER systems have been developed to support real-world applicat...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6514576/ https://www.ncbi.nlm.nih.gov/pubmed/31003522 http://dx.doi.org/10.3390/s19081863 |
_version_ | 1783417900685590528 |
---|---|
author | Samadiani, Najmeh Huang, Guangyan Cai, Borui Luo, Wei Chi, Chi-Hung Xiang, Yong He, Jing |
author_facet | Samadiani, Najmeh Huang, Guangyan Cai, Borui Luo, Wei Chi, Chi-Hung Xiang, Yong He, Jing |
author_sort | Samadiani, Najmeh |
collection | PubMed |
description | Facial Expression Recognition (FER) can be widely applied to various research areas, such as mental diseases diagnosis and human social/physiological interaction detection. With the emerging advanced technologies in hardware and sensors, FER systems have been developed to support real-world application scenes, instead of laboratory environments. Although the laboratory-controlled FER systems achieve very high accuracy, around 97%, the technical transferring from the laboratory to real-world applications faces a great barrier of very low accuracy, approximately 50%. In this survey, we comprehensively discuss three significant challenges in the unconstrained real-world environments, such as illumination variation, head pose, and subject-dependence, which may not be resolved by only analysing images/videos in the FER system. We focus on those sensors that may provide extra information and help the FER systems to detect emotion in both static images and video sequences. We introduce three categories of sensors that may help improve the accuracy and reliability of an expression recognition system by tackling the challenges mentioned above in pure image/video processing. The first group is detailed-face sensors, which detect a small dynamic change of a face component, such as eye-trackers, which may help differentiate the background noise and the feature of faces. The second is non-visual sensors, such as audio, depth, and EEG sensors, which provide extra information in addition to visual dimension and improve the recognition reliability for example in illumination variation and position shift situation. The last is target-focused sensors, such as infrared thermal sensors, which can facilitate the FER systems to filter useless visual contents and may help resist illumination variation. Also, we discuss the methods of fusing different inputs obtained from multimodal sensors in an emotion system. We comparatively review the most prominent multimodal emotional expression recognition approaches and point out their advantages and limitations. We briefly introduce the benchmark data sets related to FER systems for each category of sensors and extend our survey to the open challenges and issues. Meanwhile, we design a framework of an expression recognition system, which uses multimodal sensor data (provided by the three categories of sensors) to provide complete information about emotions to assist the pure face image/video analysis. We theoretically analyse the feasibility and achievability of our new expression recognition system, especially for the use in the wild environment, and point out the future directions to design an efficient, emotional expression recognition system. |
format | Online Article Text |
id | pubmed-6514576 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-65145762019-05-30 A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data Samadiani, Najmeh Huang, Guangyan Cai, Borui Luo, Wei Chi, Chi-Hung Xiang, Yong He, Jing Sensors (Basel) Review Facial Expression Recognition (FER) can be widely applied to various research areas, such as mental diseases diagnosis and human social/physiological interaction detection. With the emerging advanced technologies in hardware and sensors, FER systems have been developed to support real-world application scenes, instead of laboratory environments. Although the laboratory-controlled FER systems achieve very high accuracy, around 97%, the technical transferring from the laboratory to real-world applications faces a great barrier of very low accuracy, approximately 50%. In this survey, we comprehensively discuss three significant challenges in the unconstrained real-world environments, such as illumination variation, head pose, and subject-dependence, which may not be resolved by only analysing images/videos in the FER system. We focus on those sensors that may provide extra information and help the FER systems to detect emotion in both static images and video sequences. We introduce three categories of sensors that may help improve the accuracy and reliability of an expression recognition system by tackling the challenges mentioned above in pure image/video processing. The first group is detailed-face sensors, which detect a small dynamic change of a face component, such as eye-trackers, which may help differentiate the background noise and the feature of faces. The second is non-visual sensors, such as audio, depth, and EEG sensors, which provide extra information in addition to visual dimension and improve the recognition reliability for example in illumination variation and position shift situation. The last is target-focused sensors, such as infrared thermal sensors, which can facilitate the FER systems to filter useless visual contents and may help resist illumination variation. Also, we discuss the methods of fusing different inputs obtained from multimodal sensors in an emotion system. We comparatively review the most prominent multimodal emotional expression recognition approaches and point out their advantages and limitations. We briefly introduce the benchmark data sets related to FER systems for each category of sensors and extend our survey to the open challenges and issues. Meanwhile, we design a framework of an expression recognition system, which uses multimodal sensor data (provided by the three categories of sensors) to provide complete information about emotions to assist the pure face image/video analysis. We theoretically analyse the feasibility and achievability of our new expression recognition system, especially for the use in the wild environment, and point out the future directions to design an efficient, emotional expression recognition system. MDPI 2019-04-18 /pmc/articles/PMC6514576/ /pubmed/31003522 http://dx.doi.org/10.3390/s19081863 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Review Samadiani, Najmeh Huang, Guangyan Cai, Borui Luo, Wei Chi, Chi-Hung Xiang, Yong He, Jing A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data |
title | A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data |
title_full | A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data |
title_fullStr | A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data |
title_full_unstemmed | A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data |
title_short | A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data |
title_sort | review on automatic facial expression recognition systems assisted by multimodal sensor data |
topic | Review |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6514576/ https://www.ncbi.nlm.nih.gov/pubmed/31003522 http://dx.doi.org/10.3390/s19081863 |
work_keys_str_mv | AT samadianinajmeh areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT huangguangyan areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT caiborui areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT luowei areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT chichihung areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT xiangyong areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT hejing areviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT samadianinajmeh reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT huangguangyan reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT caiborui reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT luowei reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT chichihung reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT xiangyong reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata AT hejing reviewonautomaticfacialexpressionrecognitionsystemsassistedbymultimodalsensordata |