Cargando…

Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image

Mainstream methods treat head pose estimation as a supervised classification/regression problem, whose performance heavily depends on the accuracy of ground-truth labels of training data. However, it is rather difficult to obtain accurate head pose labels in practice, due to the lack of effective eq...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Leyuan, Ke, Zeran, Huo, Jiao, Chen, Jingying
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961623/
https://www.ncbi.nlm.nih.gov/pubmed/33800750
http://dx.doi.org/10.3390/s21051841
_version_ 1783665302036283392
author Liu, Leyuan
Ke, Zeran
Huo, Jiao
Chen, Jingying
author_facet Liu, Leyuan
Ke, Zeran
Huo, Jiao
Chen, Jingying
author_sort Liu, Leyuan
collection PubMed
description Mainstream methods treat head pose estimation as a supervised classification/regression problem, whose performance heavily depends on the accuracy of ground-truth labels of training data. However, it is rather difficult to obtain accurate head pose labels in practice, due to the lack of effective equipment and reasonable approaches for head pose labeling. In this paper, we propose a method which does not need to be trained with head pose labels, but matches the keypoints between a reconstructed 3D face model and the 2D input image, for head pose estimation. The proposed head pose estimation method consists of two components: the 3D face reconstruction and the 3D–2D matching keypoints. At the 3D face reconstruction phase, a personalized 3D face model is reconstructed from the input head image using convolutional neural networks, which are jointly optimized by an asymmetric Euclidean loss and a keypoint loss. At the 3D–2D keypoints matching phase, an iterative optimization algorithm is proposed to match the keypoints between the reconstructed 3D face model and the 2D input image efficiently under the constraint of perspective transformation. The proposed method is extensively evaluated on five widely used head pose estimation datasets, including Pointing’04, BIWI, AFLW2000, Multi-PIE, and Pandora. The experimental results demonstrate that the proposed method achieves excellent cross-dataset performance and surpasses most of the existing state-of-the-art approaches, with average MAEs of [Formula: see text] on Pointing’04, [Formula: see text] on BIWI, [Formula: see text] on AFLW2000, [Formula: see text] on Multi-PIE, and [Formula: see text] on Pandora, although the model of the proposed method is not trained on any of these five datasets.
format Online
Article
Text
id pubmed-7961623
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-79616232021-03-17 Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image Liu, Leyuan Ke, Zeran Huo, Jiao Chen, Jingying Sensors (Basel) Article Mainstream methods treat head pose estimation as a supervised classification/regression problem, whose performance heavily depends on the accuracy of ground-truth labels of training data. However, it is rather difficult to obtain accurate head pose labels in practice, due to the lack of effective equipment and reasonable approaches for head pose labeling. In this paper, we propose a method which does not need to be trained with head pose labels, but matches the keypoints between a reconstructed 3D face model and the 2D input image, for head pose estimation. The proposed head pose estimation method consists of two components: the 3D face reconstruction and the 3D–2D matching keypoints. At the 3D face reconstruction phase, a personalized 3D face model is reconstructed from the input head image using convolutional neural networks, which are jointly optimized by an asymmetric Euclidean loss and a keypoint loss. At the 3D–2D keypoints matching phase, an iterative optimization algorithm is proposed to match the keypoints between the reconstructed 3D face model and the 2D input image efficiently under the constraint of perspective transformation. The proposed method is extensively evaluated on five widely used head pose estimation datasets, including Pointing’04, BIWI, AFLW2000, Multi-PIE, and Pandora. The experimental results demonstrate that the proposed method achieves excellent cross-dataset performance and surpasses most of the existing state-of-the-art approaches, with average MAEs of [Formula: see text] on Pointing’04, [Formula: see text] on BIWI, [Formula: see text] on AFLW2000, [Formula: see text] on Multi-PIE, and [Formula: see text] on Pandora, although the model of the proposed method is not trained on any of these five datasets. MDPI 2021-03-06 /pmc/articles/PMC7961623/ /pubmed/33800750 http://dx.doi.org/10.3390/s21051841 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Liu, Leyuan
Ke, Zeran
Huo, Jiao
Chen, Jingying
Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image
title Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image
title_full Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image
title_fullStr Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image
title_full_unstemmed Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image
title_short Head Pose Estimation through Keypoints Matching between Reconstructed 3D Face Model and 2D Image
title_sort head pose estimation through keypoints matching between reconstructed 3d face model and 2d image
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961623/
https://www.ncbi.nlm.nih.gov/pubmed/33800750
http://dx.doi.org/10.3390/s21051841
work_keys_str_mv AT liuleyuan headposeestimationthroughkeypointsmatchingbetweenreconstructed3dfacemodeland2dimage
AT kezeran headposeestimationthroughkeypointsmatchingbetweenreconstructed3dfacemodeland2dimage
AT huojiao headposeestimationthroughkeypointsmatchingbetweenreconstructed3dfacemodeland2dimage
AT chenjingying headposeestimationthroughkeypointsmatchingbetweenreconstructed3dfacemodeland2dimage