Cargando…

An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks

Aimed at improving size, weight, and power (SWaP)-constrained robotic vision-aided state estimation, we describe our unsupervised, deep convolutional-deconvolutional sensor fusion network, Multi-Hypothesis DeepEfference (MHDE). MHDE learns to intelligently combine noisy heterogeneous sensor data to...

Descripción completa

Detalles Bibliográficos
Autores principales: Shamwell, E. Jared, Nothwang, William D., Perlis, Donald
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982537/
https://www.ncbi.nlm.nih.gov/pubmed/29734687
http://dx.doi.org/10.3390/s18051427
_version_ 1783328262488850432
author Shamwell, E. Jared
Nothwang, William D.
Perlis, Donald
author_facet Shamwell, E. Jared
Nothwang, William D.
Perlis, Donald
author_sort Shamwell, E. Jared
collection PubMed
description Aimed at improving size, weight, and power (SWaP)-constrained robotic vision-aided state estimation, we describe our unsupervised, deep convolutional-deconvolutional sensor fusion network, Multi-Hypothesis DeepEfference (MHDE). MHDE learns to intelligently combine noisy heterogeneous sensor data to predict several probable hypotheses for the dense, pixel-level correspondence between a source image and an unseen target image. We show how our multi-hypothesis formulation provides increased robustness against dynamic, heteroscedastic sensor and motion noise by computing hypothesis image mappings and predictions at 76–357 Hz depending on the number of hypotheses being generated. MHDE fuses noisy, heterogeneous sensory inputs using two parallel, inter-connected architectural pathways and n (1–20 in this work) multi-hypothesis generating sub-pathways to produce n global correspondence estimates between a source and a target image. We evaluated MHDE on the KITTI Odometry dataset and benchmarked it against the vision-only DeepMatching and Deformable Spatial Pyramids algorithms and were able to demonstrate a significant runtime decrease and a performance increase compared to the next-best performing method.
format Online
Article
Text
id pubmed-5982537
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-59825372018-06-05 An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks Shamwell, E. Jared Nothwang, William D. Perlis, Donald Sensors (Basel) Article Aimed at improving size, weight, and power (SWaP)-constrained robotic vision-aided state estimation, we describe our unsupervised, deep convolutional-deconvolutional sensor fusion network, Multi-Hypothesis DeepEfference (MHDE). MHDE learns to intelligently combine noisy heterogeneous sensor data to predict several probable hypotheses for the dense, pixel-level correspondence between a source image and an unseen target image. We show how our multi-hypothesis formulation provides increased robustness against dynamic, heteroscedastic sensor and motion noise by computing hypothesis image mappings and predictions at 76–357 Hz depending on the number of hypotheses being generated. MHDE fuses noisy, heterogeneous sensory inputs using two parallel, inter-connected architectural pathways and n (1–20 in this work) multi-hypothesis generating sub-pathways to produce n global correspondence estimates between a source and a target image. We evaluated MHDE on the KITTI Odometry dataset and benchmarked it against the vision-only DeepMatching and Deformable Spatial Pyramids algorithms and were able to demonstrate a significant runtime decrease and a performance increase compared to the next-best performing method. MDPI 2018-05-04 /pmc/articles/PMC5982537/ /pubmed/29734687 http://dx.doi.org/10.3390/s18051427 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Shamwell, E. Jared
Nothwang, William D.
Perlis, Donald
An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
title An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
title_full An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
title_fullStr An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
title_full_unstemmed An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
title_short An Embodied Multi-Sensor Fusion Approach to Visual Motion Estimation Using Unsupervised Deep Networks
title_sort embodied multi-sensor fusion approach to visual motion estimation using unsupervised deep networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982537/
https://www.ncbi.nlm.nih.gov/pubmed/29734687
http://dx.doi.org/10.3390/s18051427
work_keys_str_mv AT shamwellejared anembodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks
AT nothwangwilliamd anembodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks
AT perlisdonald anembodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks
AT shamwellejared embodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks
AT nothwangwilliamd embodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks
AT perlisdonald embodiedmultisensorfusionapproachtovisualmotionestimationusingunsuperviseddeepnetworks