Cargando…

Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors

Neuromorphic vision sensors are bio-inspired cameras that naturally capture the dynamics of a scene with ultra-low latency, filtering out redundant information with low power consumption. Few works are addressing the object detection with this sensor. In this work, we propose to develop pedestrian d...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Guang, Cao, Hu, Ye, Canbo, Zhang, Zhenyan, Liu, Xingbo, Mo, Xuhui, Qu, Zhongnan, Conradt, Jörg, Röhrbein, Florian, Knoll, Alois
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6454154/
https://www.ncbi.nlm.nih.gov/pubmed/31001104
http://dx.doi.org/10.3389/fnbot.2019.00010
_version_ 1783409517519699968
author Chen, Guang
Cao, Hu
Ye, Canbo
Zhang, Zhenyan
Liu, Xingbo
Mo, Xuhui
Qu, Zhongnan
Conradt, Jörg
Röhrbein, Florian
Knoll, Alois
author_facet Chen, Guang
Cao, Hu
Ye, Canbo
Zhang, Zhenyan
Liu, Xingbo
Mo, Xuhui
Qu, Zhongnan
Conradt, Jörg
Röhrbein, Florian
Knoll, Alois
author_sort Chen, Guang
collection PubMed
description Neuromorphic vision sensors are bio-inspired cameras that naturally capture the dynamics of a scene with ultra-low latency, filtering out redundant information with low power consumption. Few works are addressing the object detection with this sensor. In this work, we propose to develop pedestrian detectors that unlock the potential of the event data by leveraging multi-cue information and different fusion strategies. To make the best out of the event data, we introduce three different event-stream encoding methods based on Frequency, Surface of Active Event (SAE) and Leaky Integrate-and-Fire (LIF). We further integrate them into the state-of-the-art neural network architectures with two fusion approaches: the channel-level fusion of the raw feature space and decision-level fusion with the probability assignments. We present a qualitative and quantitative explanation why different encoding methods are chosen to evaluate the pedestrian detection and which method performs the best. We demonstrate the advantages of the decision-level fusion via leveraging multi-cue event information and show that our approach performs well on a self-annotated event-based pedestrian dataset with 8,736 event frames. This work paves the way of more fascinating perception applications with neuromorphic vision sensors.
format Online
Article
Text
id pubmed-6454154
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-64541542019-04-18 Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors Chen, Guang Cao, Hu Ye, Canbo Zhang, Zhenyan Liu, Xingbo Mo, Xuhui Qu, Zhongnan Conradt, Jörg Röhrbein, Florian Knoll, Alois Front Neurorobot Neuroscience Neuromorphic vision sensors are bio-inspired cameras that naturally capture the dynamics of a scene with ultra-low latency, filtering out redundant information with low power consumption. Few works are addressing the object detection with this sensor. In this work, we propose to develop pedestrian detectors that unlock the potential of the event data by leveraging multi-cue information and different fusion strategies. To make the best out of the event data, we introduce three different event-stream encoding methods based on Frequency, Surface of Active Event (SAE) and Leaky Integrate-and-Fire (LIF). We further integrate them into the state-of-the-art neural network architectures with two fusion approaches: the channel-level fusion of the raw feature space and decision-level fusion with the probability assignments. We present a qualitative and quantitative explanation why different encoding methods are chosen to evaluate the pedestrian detection and which method performs the best. We demonstrate the advantages of the decision-level fusion via leveraging multi-cue event information and show that our approach performs well on a self-annotated event-based pedestrian dataset with 8,736 event frames. This work paves the way of more fascinating perception applications with neuromorphic vision sensors. Frontiers Media S.A. 2019-04-02 /pmc/articles/PMC6454154/ /pubmed/31001104 http://dx.doi.org/10.3389/fnbot.2019.00010 Text en Copyright © 2019 Chen, Cao, Ye, Zhang, Liu, Mo, Qu, Conradt, Röhrbein and Knoll. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Chen, Guang
Cao, Hu
Ye, Canbo
Zhang, Zhenyan
Liu, Xingbo
Mo, Xuhui
Qu, Zhongnan
Conradt, Jörg
Röhrbein, Florian
Knoll, Alois
Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors
title Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors
title_full Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors
title_fullStr Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors
title_full_unstemmed Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors
title_short Multi-Cue Event Information Fusion for Pedestrian Detection With Neuromorphic Vision Sensors
title_sort multi-cue event information fusion for pedestrian detection with neuromorphic vision sensors
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6454154/
https://www.ncbi.nlm.nih.gov/pubmed/31001104
http://dx.doi.org/10.3389/fnbot.2019.00010
work_keys_str_mv AT chenguang multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT caohu multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT yecanbo multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT zhangzhenyan multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT liuxingbo multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT moxuhui multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT quzhongnan multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT conradtjorg multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT rohrbeinflorian multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors
AT knollalois multicueeventinformationfusionforpedestriandetectionwithneuromorphicvisionsensors