Cargando…

Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation

Recently, the use of convolutional neural networks for hand pose estimation from RGB images has dramatically improved. However, self-occluded keypoint inference in hand pose estimation is still a challenging task. We argue that these occluded keypoints cannot be readily recognized directly from trad...

Descripción completa

Detalles Bibliográficos
Autores principales: Guan, Xin, Shen, Huan, Nyatega, Charles Okanda, Li, Qiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10216948/
https://www.ncbi.nlm.nih.gov/pubmed/37238479
http://dx.doi.org/10.3390/e25050724
_version_ 1785048418439135232
author Guan, Xin
Shen, Huan
Nyatega, Charles Okanda
Li, Qiang
author_facet Guan, Xin
Shen, Huan
Nyatega, Charles Okanda
Li, Qiang
author_sort Guan, Xin
collection PubMed
description Recently, the use of convolutional neural networks for hand pose estimation from RGB images has dramatically improved. However, self-occluded keypoint inference in hand pose estimation is still a challenging task. We argue that these occluded keypoints cannot be readily recognized directly from traditional appearance features, and sufficient contextual information among the keypoints is especially needed to induce feature learning. Therefore, we propose a new repeated cross-scale structure-induced feature fusion network to learn about the representations of keypoints with rich information, ’informed’ by the relationships between different abstraction levels of features. Our network consists of two modules: GlobalNet and RegionalNet. GlobalNet roughly locates hand joints based on a new feature pyramid structure by combining higher semantic information and more global spatial scale information. RegionalNet further refines keypoint representation learning via a four-stage cross-scale feature fusion network, which learns shallow appearance features induced by more implicit hand structure information, so that when identifying occluded keypoints, the network can use augmented features to better locate the positions. The experimental results show that our method outperforms the state-of-the-art methods for 2D hand pose estimation on two public datasets, STB and RHD.
format Online
Article
Text
id pubmed-10216948
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102169482023-05-27 Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation Guan, Xin Shen, Huan Nyatega, Charles Okanda Li, Qiang Entropy (Basel) Article Recently, the use of convolutional neural networks for hand pose estimation from RGB images has dramatically improved. However, self-occluded keypoint inference in hand pose estimation is still a challenging task. We argue that these occluded keypoints cannot be readily recognized directly from traditional appearance features, and sufficient contextual information among the keypoints is especially needed to induce feature learning. Therefore, we propose a new repeated cross-scale structure-induced feature fusion network to learn about the representations of keypoints with rich information, ’informed’ by the relationships between different abstraction levels of features. Our network consists of two modules: GlobalNet and RegionalNet. GlobalNet roughly locates hand joints based on a new feature pyramid structure by combining higher semantic information and more global spatial scale information. RegionalNet further refines keypoint representation learning via a four-stage cross-scale feature fusion network, which learns shallow appearance features induced by more implicit hand structure information, so that when identifying occluded keypoints, the network can use augmented features to better locate the positions. The experimental results show that our method outperforms the state-of-the-art methods for 2D hand pose estimation on two public datasets, STB and RHD. MDPI 2023-04-27 /pmc/articles/PMC10216948/ /pubmed/37238479 http://dx.doi.org/10.3390/e25050724 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Guan, Xin
Shen, Huan
Nyatega, Charles Okanda
Li, Qiang
Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation
title Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation
title_full Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation
title_fullStr Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation
title_full_unstemmed Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation
title_short Repeated Cross-Scale Structure-Induced Feature Fusion Network for 2D Hand Pose Estimation
title_sort repeated cross-scale structure-induced feature fusion network for 2d hand pose estimation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10216948/
https://www.ncbi.nlm.nih.gov/pubmed/37238479
http://dx.doi.org/10.3390/e25050724
work_keys_str_mv AT guanxin repeatedcrossscalestructureinducedfeaturefusionnetworkfor2dhandposeestimation
AT shenhuan repeatedcrossscalestructureinducedfeaturefusionnetworkfor2dhandposeestimation
AT nyategacharlesokanda repeatedcrossscalestructureinducedfeaturefusionnetworkfor2dhandposeestimation
AT liqiang repeatedcrossscalestructureinducedfeaturefusionnetworkfor2dhandposeestimation