Cargando…
BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference
Edge-cloud collaborative inference can significantly reduce the delay of a deep neural network (DNN) by dividing the network between mobile edge and cloud. However, the in-layer data size of DNN is usually larger than the original data, so the communication time to send intermediate data to the clou...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8272083/ https://www.ncbi.nlm.nih.gov/pubmed/34209400 http://dx.doi.org/10.3390/s21134494 |
_version_ | 1783721141936848896 |
---|---|
author | Zhou, Hongbo Zhang, Weiwei Wang, Chengwei Ma, Xin Yu, Haoran |
author_facet | Zhou, Hongbo Zhang, Weiwei Wang, Chengwei Ma, Xin Yu, Haoran |
author_sort | Zhou, Hongbo |
collection | PubMed |
description | Edge-cloud collaborative inference can significantly reduce the delay of a deep neural network (DNN) by dividing the network between mobile edge and cloud. However, the in-layer data size of DNN is usually larger than the original data, so the communication time to send intermediate data to the cloud will also increase end-to-end latency. To cope with these challenges, this paper proposes a novel convolutional neural network structure—BBNet—that accelerates collaborative inference from two levels: (1) through channel-pruning: reducing the number of calculations and parameters of the original network; (2) through compressing the feature map at the split point to further reduce the size of the data transmitted. In addition, This paper implemented the BBNet structure based on NVIDIA Nano and the server. Compared with the original network, BBNet’s FLOPs and parameter achieve up to 5.67× and 11.57× on the compression rate, respectively. In the best case, the feature compression layer can reach a bit-compression rate of 512×. Compared with the better bandwidth conditions, BBNet has a more obvious inference delay when the network conditions are poor. For example, when the upload bandwidth is only 20 kb/s, the end-to-end latency of BBNet is increased by 38.89× compared with the cloud-only approach. |
format | Online Article Text |
id | pubmed-8272083 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-82720832021-07-11 BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference Zhou, Hongbo Zhang, Weiwei Wang, Chengwei Ma, Xin Yu, Haoran Sensors (Basel) Article Edge-cloud collaborative inference can significantly reduce the delay of a deep neural network (DNN) by dividing the network between mobile edge and cloud. However, the in-layer data size of DNN is usually larger than the original data, so the communication time to send intermediate data to the cloud will also increase end-to-end latency. To cope with these challenges, this paper proposes a novel convolutional neural network structure—BBNet—that accelerates collaborative inference from two levels: (1) through channel-pruning: reducing the number of calculations and parameters of the original network; (2) through compressing the feature map at the split point to further reduce the size of the data transmitted. In addition, This paper implemented the BBNet structure based on NVIDIA Nano and the server. Compared with the original network, BBNet’s FLOPs and parameter achieve up to 5.67× and 11.57× on the compression rate, respectively. In the best case, the feature compression layer can reach a bit-compression rate of 512×. Compared with the better bandwidth conditions, BBNet has a more obvious inference delay when the network conditions are poor. For example, when the upload bandwidth is only 20 kb/s, the end-to-end latency of BBNet is increased by 38.89× compared with the cloud-only approach. MDPI 2021-06-30 /pmc/articles/PMC8272083/ /pubmed/34209400 http://dx.doi.org/10.3390/s21134494 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhou, Hongbo Zhang, Weiwei Wang, Chengwei Ma, Xin Yu, Haoran BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference |
title | BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference |
title_full | BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference |
title_fullStr | BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference |
title_full_unstemmed | BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference |
title_short | BBNet: A Novel Convolutional Neural Network Structure in Edge-Cloud Collaborative Inference |
title_sort | bbnet: a novel convolutional neural network structure in edge-cloud collaborative inference |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8272083/ https://www.ncbi.nlm.nih.gov/pubmed/34209400 http://dx.doi.org/10.3390/s21134494 |
work_keys_str_mv | AT zhouhongbo bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference AT zhangweiwei bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference AT wangchengwei bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference AT maxin bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference AT yuhaoran bbnetanovelconvolutionalneuralnetworkstructureinedgecloudcollaborativeinference |