Cargando…

An Occlusion-Aware Framework for Real-Time 3D Pose Tracking

Random forest-based methods for 3D temporal tracking over an image sequence have gained increasing prominence in recent years. They do not require object’s texture and only use the raw depth images and previous pose as input, which makes them especially suitable for textureless objects. These method...

Descripción completa

Detalles Bibliográficos
Autores principales: Fu, Mingliang, Leng, Yuquan, Luo, Haitao, Zhou, Weijia
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111635/
https://www.ncbi.nlm.nih.gov/pubmed/30127294
http://dx.doi.org/10.3390/s18082734
_version_ 1783350695650394112
author Fu, Mingliang
Leng, Yuquan
Luo, Haitao
Zhou, Weijia
author_facet Fu, Mingliang
Leng, Yuquan
Luo, Haitao
Zhou, Weijia
author_sort Fu, Mingliang
collection PubMed
description Random forest-based methods for 3D temporal tracking over an image sequence have gained increasing prominence in recent years. They do not require object’s texture and only use the raw depth images and previous pose as input, which makes them especially suitable for textureless objects. These methods learn a built-in occlusion handling from predetermined occlusion patterns, which are not always able to model the real case. Besides, the input of random forest is mixed with more and more outliers as the occlusion deepens. In this paper, we propose an occlusion-aware framework capable of real-time and robust 3D pose tracking from RGB-D images. To this end, the proposed framework is anchored in the random forest-based learning strategy, referred to as RFtracker. We aim to enhance its performance from two aspects: integrated local refinement of random forest on one side, and online rendering based occlusion handling on the other. In order to eliminate the inconsistency between learning and prediction of RFtracker, a local refinement step is embedded to guide random forest towards the optimal regression. Furthermore, we present an online rendering-based occlusion handling to improve the robustness against dynamic occlusion. Meanwhile, a lightweight convolutional neural network-based motion-compensated (CMC) module is designed to cope with fast motion and inevitable physical delay caused by imaging frequency and data transmission. Finally, experiments show that our proposed framework can cope better with heavily-occluded scenes than RFtracker and preserve the real-time performance.
format Online
Article
Text
id pubmed-6111635
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-61116352018-08-30 An Occlusion-Aware Framework for Real-Time 3D Pose Tracking Fu, Mingliang Leng, Yuquan Luo, Haitao Zhou, Weijia Sensors (Basel) Article Random forest-based methods for 3D temporal tracking over an image sequence have gained increasing prominence in recent years. They do not require object’s texture and only use the raw depth images and previous pose as input, which makes them especially suitable for textureless objects. These methods learn a built-in occlusion handling from predetermined occlusion patterns, which are not always able to model the real case. Besides, the input of random forest is mixed with more and more outliers as the occlusion deepens. In this paper, we propose an occlusion-aware framework capable of real-time and robust 3D pose tracking from RGB-D images. To this end, the proposed framework is anchored in the random forest-based learning strategy, referred to as RFtracker. We aim to enhance its performance from two aspects: integrated local refinement of random forest on one side, and online rendering based occlusion handling on the other. In order to eliminate the inconsistency between learning and prediction of RFtracker, a local refinement step is embedded to guide random forest towards the optimal regression. Furthermore, we present an online rendering-based occlusion handling to improve the robustness against dynamic occlusion. Meanwhile, a lightweight convolutional neural network-based motion-compensated (CMC) module is designed to cope with fast motion and inevitable physical delay caused by imaging frequency and data transmission. Finally, experiments show that our proposed framework can cope better with heavily-occluded scenes than RFtracker and preserve the real-time performance. MDPI 2018-08-20 /pmc/articles/PMC6111635/ /pubmed/30127294 http://dx.doi.org/10.3390/s18082734 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Fu, Mingliang
Leng, Yuquan
Luo, Haitao
Zhou, Weijia
An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
title An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
title_full An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
title_fullStr An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
title_full_unstemmed An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
title_short An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
title_sort occlusion-aware framework for real-time 3d pose tracking
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111635/
https://www.ncbi.nlm.nih.gov/pubmed/30127294
http://dx.doi.org/10.3390/s18082734
work_keys_str_mv AT fumingliang anocclusionawareframeworkforrealtime3dposetracking
AT lengyuquan anocclusionawareframeworkforrealtime3dposetracking
AT luohaitao anocclusionawareframeworkforrealtime3dposetracking
AT zhouweijia anocclusionawareframeworkforrealtime3dposetracking
AT fumingliang occlusionawareframeworkforrealtime3dposetracking
AT lengyuquan occlusionawareframeworkforrealtime3dposetracking
AT luohaitao occlusionawareframeworkforrealtime3dposetracking
AT zhouweijia occlusionawareframeworkforrealtime3dposetracking