Cargando…

Real Time 3D Facial Movement Tracking Using a Monocular Camera

The paper proposes a robust framework for 3D facial movement tracking in real time using a monocular camera. It is designed to estimate the 3D face pose and local facial animation such as eyelid movement and mouth movement. The framework firstly utilizes the Discriminative Shape Regression method to...

Descripción completa

Detalles Bibliográficos
Autores principales: Dong, Yanchao, Wang, Yanming, Yue, Jiguang, Hu, Zhencheng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5017323/
https://www.ncbi.nlm.nih.gov/pubmed/27463714
http://dx.doi.org/10.3390/s16081157
_version_ 1782452721632149504
author Dong, Yanchao
Wang, Yanming
Yue, Jiguang
Hu, Zhencheng
author_facet Dong, Yanchao
Wang, Yanming
Yue, Jiguang
Hu, Zhencheng
author_sort Dong, Yanchao
collection PubMed
description The paper proposes a robust framework for 3D facial movement tracking in real time using a monocular camera. It is designed to estimate the 3D face pose and local facial animation such as eyelid movement and mouth movement. The framework firstly utilizes the Discriminative Shape Regression method to locate the facial feature points on the 2D image and fuses the 2D data with a 3D face model using Extended Kalman Filter to yield 3D facial movement information. An alternating optimizing strategy is adopted to fit to different persons automatically. Experiments show that the proposed framework could track the 3D facial movement across various poses and illumination conditions. Given the real face scale the framework could track the eyelid with an error of 1 mm and mouth with an error of 2 mm. The tracking result is reliable for expression analysis or mental state inference.
format Online
Article
Text
id pubmed-5017323
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-50173232016-09-22 Real Time 3D Facial Movement Tracking Using a Monocular Camera Dong, Yanchao Wang, Yanming Yue, Jiguang Hu, Zhencheng Sensors (Basel) Article The paper proposes a robust framework for 3D facial movement tracking in real time using a monocular camera. It is designed to estimate the 3D face pose and local facial animation such as eyelid movement and mouth movement. The framework firstly utilizes the Discriminative Shape Regression method to locate the facial feature points on the 2D image and fuses the 2D data with a 3D face model using Extended Kalman Filter to yield 3D facial movement information. An alternating optimizing strategy is adopted to fit to different persons automatically. Experiments show that the proposed framework could track the 3D facial movement across various poses and illumination conditions. Given the real face scale the framework could track the eyelid with an error of 1 mm and mouth with an error of 2 mm. The tracking result is reliable for expression analysis or mental state inference. MDPI 2016-07-25 /pmc/articles/PMC5017323/ /pubmed/27463714 http://dx.doi.org/10.3390/s16081157 Text en © 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Dong, Yanchao
Wang, Yanming
Yue, Jiguang
Hu, Zhencheng
Real Time 3D Facial Movement Tracking Using a Monocular Camera
title Real Time 3D Facial Movement Tracking Using a Monocular Camera
title_full Real Time 3D Facial Movement Tracking Using a Monocular Camera
title_fullStr Real Time 3D Facial Movement Tracking Using a Monocular Camera
title_full_unstemmed Real Time 3D Facial Movement Tracking Using a Monocular Camera
title_short Real Time 3D Facial Movement Tracking Using a Monocular Camera
title_sort real time 3d facial movement tracking using a monocular camera
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5017323/
https://www.ncbi.nlm.nih.gov/pubmed/27463714
http://dx.doi.org/10.3390/s16081157
work_keys_str_mv AT dongyanchao realtime3dfacialmovementtrackingusingamonocularcamera
AT wangyanming realtime3dfacialmovementtrackingusingamonocularcamera
AT yuejiguang realtime3dfacialmovementtrackingusingamonocularcamera
AT huzhencheng realtime3dfacialmovementtrackingusingamonocularcamera