Cargando…
An Occlusion-Aware Framework for Real-Time 3D Pose Tracking
Random forest-based methods for 3D temporal tracking over an image sequence have gained increasing prominence in recent years. They do not require object’s texture and only use the raw depth images and previous pose as input, which makes them especially suitable for textureless objects. These method...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111635/ https://www.ncbi.nlm.nih.gov/pubmed/30127294 http://dx.doi.org/10.3390/s18082734 |
_version_ | 1783350695650394112 |
---|---|
author | Fu, Mingliang Leng, Yuquan Luo, Haitao Zhou, Weijia |
author_facet | Fu, Mingliang Leng, Yuquan Luo, Haitao Zhou, Weijia |
author_sort | Fu, Mingliang |
collection | PubMed |
description | Random forest-based methods for 3D temporal tracking over an image sequence have gained increasing prominence in recent years. They do not require object’s texture and only use the raw depth images and previous pose as input, which makes them especially suitable for textureless objects. These methods learn a built-in occlusion handling from predetermined occlusion patterns, which are not always able to model the real case. Besides, the input of random forest is mixed with more and more outliers as the occlusion deepens. In this paper, we propose an occlusion-aware framework capable of real-time and robust 3D pose tracking from RGB-D images. To this end, the proposed framework is anchored in the random forest-based learning strategy, referred to as RFtracker. We aim to enhance its performance from two aspects: integrated local refinement of random forest on one side, and online rendering based occlusion handling on the other. In order to eliminate the inconsistency between learning and prediction of RFtracker, a local refinement step is embedded to guide random forest towards the optimal regression. Furthermore, we present an online rendering-based occlusion handling to improve the robustness against dynamic occlusion. Meanwhile, a lightweight convolutional neural network-based motion-compensated (CMC) module is designed to cope with fast motion and inevitable physical delay caused by imaging frequency and data transmission. Finally, experiments show that our proposed framework can cope better with heavily-occluded scenes than RFtracker and preserve the real-time performance. |
format | Online Article Text |
id | pubmed-6111635 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-61116352018-08-30 An Occlusion-Aware Framework for Real-Time 3D Pose Tracking Fu, Mingliang Leng, Yuquan Luo, Haitao Zhou, Weijia Sensors (Basel) Article Random forest-based methods for 3D temporal tracking over an image sequence have gained increasing prominence in recent years. They do not require object’s texture and only use the raw depth images and previous pose as input, which makes them especially suitable for textureless objects. These methods learn a built-in occlusion handling from predetermined occlusion patterns, which are not always able to model the real case. Besides, the input of random forest is mixed with more and more outliers as the occlusion deepens. In this paper, we propose an occlusion-aware framework capable of real-time and robust 3D pose tracking from RGB-D images. To this end, the proposed framework is anchored in the random forest-based learning strategy, referred to as RFtracker. We aim to enhance its performance from two aspects: integrated local refinement of random forest on one side, and online rendering based occlusion handling on the other. In order to eliminate the inconsistency between learning and prediction of RFtracker, a local refinement step is embedded to guide random forest towards the optimal regression. Furthermore, we present an online rendering-based occlusion handling to improve the robustness against dynamic occlusion. Meanwhile, a lightweight convolutional neural network-based motion-compensated (CMC) module is designed to cope with fast motion and inevitable physical delay caused by imaging frequency and data transmission. Finally, experiments show that our proposed framework can cope better with heavily-occluded scenes than RFtracker and preserve the real-time performance. MDPI 2018-08-20 /pmc/articles/PMC6111635/ /pubmed/30127294 http://dx.doi.org/10.3390/s18082734 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Fu, Mingliang Leng, Yuquan Luo, Haitao Zhou, Weijia An Occlusion-Aware Framework for Real-Time 3D Pose Tracking |
title | An Occlusion-Aware Framework for Real-Time 3D Pose Tracking |
title_full | An Occlusion-Aware Framework for Real-Time 3D Pose Tracking |
title_fullStr | An Occlusion-Aware Framework for Real-Time 3D Pose Tracking |
title_full_unstemmed | An Occlusion-Aware Framework for Real-Time 3D Pose Tracking |
title_short | An Occlusion-Aware Framework for Real-Time 3D Pose Tracking |
title_sort | occlusion-aware framework for real-time 3d pose tracking |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111635/ https://www.ncbi.nlm.nih.gov/pubmed/30127294 http://dx.doi.org/10.3390/s18082734 |
work_keys_str_mv | AT fumingliang anocclusionawareframeworkforrealtime3dposetracking AT lengyuquan anocclusionawareframeworkforrealtime3dposetracking AT luohaitao anocclusionawareframeworkforrealtime3dposetracking AT zhouweijia anocclusionawareframeworkforrealtime3dposetracking AT fumingliang occlusionawareframeworkforrealtime3dposetracking AT lengyuquan occlusionawareframeworkforrealtime3dposetracking AT luohaitao occlusionawareframeworkforrealtime3dposetracking AT zhouweijia occlusionawareframeworkforrealtime3dposetracking |