Cargando…

BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference

Edge-cloud collaborative inference can significantly reduce the delay of a deep neural network (DNN) by dividing the network between mobile edge and cloud. However, the in-layer data size of DNN is usually larger than the original data, so the communication time to send intermediate data to the clou...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhou, Hongbo, Zhang, Weiwei, Wang, Chengwei, Ma, Xin, Yu, Haoran
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8272083/
https://www.ncbi.nlm.nih.gov/pubmed/34209400
http://dx.doi.org/10.3390/s21134494
_version_ 1783721141936848896
author Zhou, Hongbo
Zhang, Weiwei
Wang, Chengwei
Ma, Xin
Yu, Haoran
author_facet Zhou, Hongbo
Zhang, Weiwei
Wang, Chengwei
Ma, Xin
Yu, Haoran
author_sort Zhou, Hongbo
collection PubMed
description Edge-cloud collaborative inference can significantly reduce the delay of a deep neural network (DNN) by dividing the network between mobile edge and cloud. However, the in-layer data size of DNN is usually larger than the original data, so the communication time to send intermediate data to the cloud will also increase end-to-end latency. To cope with these challenges, this paper proposes a novel convolutional neural network structure—BBNet—that accelerates collaborative inference from two levels: (1) through channel-pruning: reducing the number of calculations and parameters of the original network; (2) through compressing the feature map at the split point to further reduce the size of the data transmitted. In addition, This paper implemented the BBNet structure based on NVIDIA Nano and the server. Compared with the original network, BBNet’s FLOPs and parameter achieve up to 5.67× and 11.57× on the compression rate, respectively. In the best case, the feature compression layer can reach a bit-compression rate of 512×. Compared with the better bandwidth conditions, BBNet has a more obvious inference delay when the network conditions are poor. For example, when the upload bandwidth is only 20 kb/s, the end-to-end latency of BBNet is increased by 38.89× compared with the cloud-only approach.
format Online
Article
Text
id pubmed-8272083
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-82720832021-07-11 BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference Zhou, Hongbo Zhang, Weiwei Wang, Chengwei Ma, Xin Yu, Haoran Sensors (Basel) Article Edge-cloud collaborative inference can significantly reduce the delay of a deep neural network (DNN) by dividing the network between mobile edge and cloud. However, the in-layer data size of DNN is usually larger than the original data, so the communication time to send intermediate data to the cloud will also increase end-to-end latency. To cope with these challenges, this paper proposes a novel convolutional neural network structure—BBNet—that accelerates collaborative inference from two levels: (1) through channel-pruning: reducing the number of calculations and parameters of the original network; (2) through compressing the feature map at the split point to further reduce the size of the data transmitted. In addition, This paper implemented the BBNet structure based on NVIDIA Nano and the server. Compared with the original network, BBNet’s FLOPs and parameter achieve up to 5.67× and 11.57× on the compression rate, respectively. In the best case, the feature compression layer can reach a bit-compression rate of 512×. Compared with the better bandwidth conditions, BBNet has a more obvious inference delay when the network conditions are poor. For example, when the upload bandwidth is only 20 kb/s, the end-to-end latency of BBNet is increased by 38.89× compared with the cloud-only approach. MDPI 2021-06-30 /pmc/articles/PMC8272083/ /pubmed/34209400 http://dx.doi.org/10.3390/s21134494 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhou, Hongbo
Zhang, Weiwei
Wang, Chengwei
Ma, Xin
Yu, Haoran
BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
title BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
title_full BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
title_fullStr BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
title_full_unstemmed BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
title_short BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
title_sort bbnet: a novel convolutional neural network structure in edge-cloud collaborative inference
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8272083/
https://www.ncbi.nlm.nih.gov/pubmed/34209400
http://dx.doi.org/10.3390/s21134494
work_keys_str_mv AT zhouhongbo bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference
AT zhangweiwei bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference
AT wangchengwei bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference
AT maxin bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference
AT yuhaoran bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference