Cargando…
Energy-Efficient Spiking Segmenter for Frame and Event-Based Images
Semantic segmentation predicts dense pixel-wise semantic labels, which is crucial for autonomous environment perception systems. For applications on mobile devices, current research focuses on energy-efficient segmenters for both frame and event-based cameras. However, there is currently no artifici...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10452323/ https://www.ncbi.nlm.nih.gov/pubmed/37622961 http://dx.doi.org/10.3390/biomimetics8040356 |
_version_ | 1785095640131305472 |
---|---|
author | Zhang, Hong Fan, Xiongfei Zhang, Yu |
author_facet | Zhang, Hong Fan, Xiongfei Zhang, Yu |
author_sort | Zhang, Hong |
collection | PubMed |
description | Semantic segmentation predicts dense pixel-wise semantic labels, which is crucial for autonomous environment perception systems. For applications on mobile devices, current research focuses on energy-efficient segmenters for both frame and event-based cameras. However, there is currently no artificial neural network (ANN) that can perform efficient segmentation on both types of images. This paper introduces spiking neural network (SNN, a bionic model that is energy-efficient when implemented on neuromorphic hardware) and develops a Spiking Context Guided Network (Spiking CGNet) with substantially lower energy consumption and comparable performance for both frame and event-based images. First, this paper proposes a spiking context guided block that can extract local features and context information with spike computations. On this basis, the directly-trained SCGNet-S and SCGNet-L are established for both frame and event-based images. Our method is verified on the frame-based dataset Cityscapes and the event-based dataset DDD17. On the Cityscapes dataset, SCGNet-S achieves comparable results to ANN CGNet with 4.85 × energy efficiency. On the DDD17 dataset, Spiking CGNet outperforms other spiking segmenters by a large margin. |
format | Online Article Text |
id | pubmed-10452323 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-104523232023-08-26 Energy-Efficient Spiking Segmenter for Frame and Event-Based Images Zhang, Hong Fan, Xiongfei Zhang, Yu Biomimetics (Basel) Article Semantic segmentation predicts dense pixel-wise semantic labels, which is crucial for autonomous environment perception systems. For applications on mobile devices, current research focuses on energy-efficient segmenters for both frame and event-based cameras. However, there is currently no artificial neural network (ANN) that can perform efficient segmentation on both types of images. This paper introduces spiking neural network (SNN, a bionic model that is energy-efficient when implemented on neuromorphic hardware) and develops a Spiking Context Guided Network (Spiking CGNet) with substantially lower energy consumption and comparable performance for both frame and event-based images. First, this paper proposes a spiking context guided block that can extract local features and context information with spike computations. On this basis, the directly-trained SCGNet-S and SCGNet-L are established for both frame and event-based images. Our method is verified on the frame-based dataset Cityscapes and the event-based dataset DDD17. On the Cityscapes dataset, SCGNet-S achieves comparable results to ANN CGNet with 4.85 × energy efficiency. On the DDD17 dataset, Spiking CGNet outperforms other spiking segmenters by a large margin. MDPI 2023-08-10 /pmc/articles/PMC10452323/ /pubmed/37622961 http://dx.doi.org/10.3390/biomimetics8040356 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhang, Hong Fan, Xiongfei Zhang, Yu Energy-Efficient Spiking Segmenter for Frame and Event-Based Images |
title | Energy-Efficient Spiking Segmenter for Frame and Event-Based Images |
title_full | Energy-Efficient Spiking Segmenter for Frame and Event-Based Images |
title_fullStr | Energy-Efficient Spiking Segmenter for Frame and Event-Based Images |
title_full_unstemmed | Energy-Efficient Spiking Segmenter for Frame and Event-Based Images |
title_short | Energy-Efficient Spiking Segmenter for Frame and Event-Based Images |
title_sort | energy-efficient spiking segmenter for frame and event-based images |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10452323/ https://www.ncbi.nlm.nih.gov/pubmed/37622961 http://dx.doi.org/10.3390/biomimetics8040356 |
work_keys_str_mv | AT zhanghong energyefficientspikingsegmenterforframeandeventbasedimages AT fanxiongfei energyefficientspikingsegmenterforframeandeventbasedimages AT zhangyu energyefficientspikingsegmenterforframeandeventbasedimages |