Cargando…
An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
Aimed at improving size, weight, and power (SWaP)-constrained robotic vision-aided state estimation, we describe our unsupervised, deep convolutional-deconvolutional sensor fusion network, Multi-Hypothesis DeepEfference (MHDE). MHDE learns to intelligently combine noisy heterogeneous sensor data to...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982537/ https://www.ncbi.nlm.nih.gov/pubmed/29734687 http://dx.doi.org/10.3390/s18051427 |
_version_ | 1783328262488850432 |
---|---|
author | Shamwell, E. Jared Nothwang, William D. Perlis, Donald |
author_facet | Shamwell, E. Jared Nothwang, William D. Perlis, Donald |
author_sort | Shamwell, E. Jared |
collection | PubMed |
description | Aimed at improving size, weight, and power (SWaP)-constrained robotic vision-aided state estimation, we describe our unsupervised, deep convolutional-deconvolutional sensor fusion network, Multi-Hypothesis DeepEfference (MHDE). MHDE learns to intelligently combine noisy heterogeneous sensor data to predict several probable hypotheses for the dense, pixel-level correspondence between a source image and an unseen target image. We show how our multi-hypothesis formulation provides increased robustness against dynamic, heteroscedastic sensor and motion noise by computing hypothesis image mappings and predictions at 76–357 Hz depending on the number of hypotheses being generated. MHDE fuses noisy, heterogeneous sensory inputs using two parallel, inter-connected architectural pathways and n (1–20 in this work) multi-hypothesis generating sub-pathways to produce n global correspondence estimates between a source and a target image. We evaluated MHDE on the KITTI Odometry dataset and benchmarked it against the vision-only DeepMatching and Deformable Spatial Pyramids algorithms and were able to demonstrate a significant runtime decrease and a performance increase compared to the next-best performing method. |
format | Online Article Text |
id | pubmed-5982537 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-59825372018-06-05 An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks Shamwell, E. Jared Nothwang, William D. Perlis, Donald Sensors (Basel) Article Aimed at improving size, weight, and power (SWaP)-constrained robotic vision-aided state estimation, we describe our unsupervised, deep convolutional-deconvolutional sensor fusion network, Multi-Hypothesis DeepEfference (MHDE). MHDE learns to intelligently combine noisy heterogeneous sensor data to predict several probable hypotheses for the dense, pixel-level correspondence between a source image and an unseen target image. We show how our multi-hypothesis formulation provides increased robustness against dynamic, heteroscedastic sensor and motion noise by computing hypothesis image mappings and predictions at 76–357 Hz depending on the number of hypotheses being generated. MHDE fuses noisy, heterogeneous sensory inputs using two parallel, inter-connected architectural pathways and n (1–20 in this work) multi-hypothesis generating sub-pathways to produce n global correspondence estimates between a source and a target image. We evaluated MHDE on the KITTI Odometry dataset and benchmarked it against the vision-only DeepMatching and Deformable Spatial Pyramids algorithms and were able to demonstrate a significant runtime decrease and a performance increase compared to the next-best performing method. MDPI 2018-05-04 /pmc/articles/PMC5982537/ /pubmed/29734687 http://dx.doi.org/10.3390/s18051427 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Shamwell, E. Jared Nothwang, William D. Perlis, Donald An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks |
title | An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks |
title_full | An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks |
title_fullStr | An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks |
title_full_unstemmed | An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks |
title_short | An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks |
title_sort | embodied multi-sensor fusion approach to visual motion estimation using unsupervised deep networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982537/ https://www.ncbi.nlm.nih.gov/pubmed/29734687 http://dx.doi.org/10.3390/s18051427 |
work_keys_str_mv | AT shamwellejared anembodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks AT nothwangwilliamd anembodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks AT perlisdonald anembodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks AT shamwellejared embodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks AT nothwangwilliamd embodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks AT perlisdonald embodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks |