Cargando…

Optimization of Edge Resources for Deep Learning Application with Batch and Model Management

As deep learning technology paves its way, real-world applications that make use of it become popular these days. Edge computing architecture is one of the service architectures to realize the deep learning based service, which makes use of the resources near the data source or client. In Edge compu...

Descripción completa

Detalles Bibliográficos
Autores principales: Kum, Seungwoo, Oh, Seungtaek, Yeom, Jeongcheol, Moon, Jaewon
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9460489/
https://www.ncbi.nlm.nih.gov/pubmed/36081177
http://dx.doi.org/10.3390/s22176717
_version_ 1784786759995883520
author Kum, Seungwoo
Oh, Seungtaek
Yeom, Jeongcheol
Moon, Jaewon
author_facet Kum, Seungwoo
Oh, Seungtaek
Yeom, Jeongcheol
Moon, Jaewon
author_sort Kum, Seungwoo
collection PubMed
description As deep learning technology paves its way, real-world applications that make use of it become popular these days. Edge computing architecture is one of the service architectures to realize the deep learning based service, which makes use of the resources near the data source or client. In Edge computing architecture it becomes important to manage resource usage, and there is research on optimization of deep learning, such as pruning or binarization, which makes deep learning models more lightweight, along with the research for the efficient distribution of workloads on cloud or edge resources. Those are to reduce the workload on edge resources. In this paper, a usage optimization method with batch and model management is proposed. The proposed method is to increase the utilization of GPU resource by modifying the batch size of the input of an inference application. To this end, the inference pipelines are identified to see how the different kinds of resources are used, and then the effect of batch inference on GPU is measured. The proposed method consists of a few modules, including a tool for batch size management which is able to change a batch size with respect to the available resources, and another one for model management which supports on-the-fly update of a model. The proposed methods are implemented on a real-time video analysis application and deployed in the Kubernetes cluster as a Docker container. The result shows that the proposed method can optimize the usage of edge resources for real-time video analysis deep learning applications.
format Online
Article
Text
id pubmed-9460489
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-94604892022-09-10 Optimization of Edge Resources for Deep Learning Application with Batch and Model Management Kum, Seungwoo Oh, Seungtaek Yeom, Jeongcheol Moon, Jaewon Sensors (Basel) Article As deep learning technology paves its way, real-world applications that make use of it become popular these days. Edge computing architecture is one of the service architectures to realize the deep learning based service, which makes use of the resources near the data source or client. In Edge computing architecture it becomes important to manage resource usage, and there is research on optimization of deep learning, such as pruning or binarization, which makes deep learning models more lightweight, along with the research for the efficient distribution of workloads on cloud or edge resources. Those are to reduce the workload on edge resources. In this paper, a usage optimization method with batch and model management is proposed. The proposed method is to increase the utilization of GPU resource by modifying the batch size of the input of an inference application. To this end, the inference pipelines are identified to see how the different kinds of resources are used, and then the effect of batch inference on GPU is measured. The proposed method consists of a few modules, including a tool for batch size management which is able to change a batch size with respect to the available resources, and another one for model management which supports on-the-fly update of a model. The proposed methods are implemented on a real-time video analysis application and deployed in the Kubernetes cluster as a Docker container. The result shows that the proposed method can optimize the usage of edge resources for real-time video analysis deep learning applications. MDPI 2022-09-05 /pmc/articles/PMC9460489/ /pubmed/36081177 http://dx.doi.org/10.3390/s22176717 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kum, Seungwoo
Oh, Seungtaek
Yeom, Jeongcheol
Moon, Jaewon
Optimization of Edge Resources for Deep Learning Application with Batch and Model Management
title Optimization of Edge Resources for Deep Learning Application with Batch and Model Management
title_full Optimization of Edge Resources for Deep Learning Application with Batch and Model Management
title_fullStr Optimization of Edge Resources for Deep Learning Application with Batch and Model Management
title_full_unstemmed Optimization of Edge Resources for Deep Learning Application with Batch and Model Management
title_short Optimization of Edge Resources for Deep Learning Application with Batch and Model Management
title_sort optimization of edge resources for deep learning application with batch and model management
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9460489/
https://www.ncbi.nlm.nih.gov/pubmed/36081177
http://dx.doi.org/10.3390/s22176717
work_keys_str_mv AT kumseungwoo optimizationofedgeresourcesfordeeplearningapplicationwithbatchandmodelmanagement
AT ohseungtaek optimizationofedgeresourcesfordeeplearningapplicationwithbatchandmodelmanagement
AT yeomjeongcheol optimizationofedgeresourcesfordeeplearningapplicationwithbatchandmodelmanagement
AT moonjaewon optimizationofedgeresourcesfordeeplearningapplicationwithbatchandmodelmanagement