Cargando…
An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
For many automotive functionalities in Advanced Driver Assist Systems (ADAS) and Autonomous Driving (AD), target objects are detected using state-of-the-art Deep Neural Network (DNN) technologies. However, the main challenge of recent DNN-based object detection is that it requires high computational...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10142959/ https://www.ncbi.nlm.nih.gov/pubmed/37112333 http://dx.doi.org/10.3390/s23083992 |
_version_ | 1785033737143058432 |
---|---|
author | Park, Jungme Aryal, Pawan Mandumula, Sai Rithvick Asolkar, Ritwik Prasad |
author_facet | Park, Jungme Aryal, Pawan Mandumula, Sai Rithvick Asolkar, Ritwik Prasad |
author_sort | Park, Jungme |
collection | PubMed |
description | For many automotive functionalities in Advanced Driver Assist Systems (ADAS) and Autonomous Driving (AD), target objects are detected using state-of-the-art Deep Neural Network (DNN) technologies. However, the main challenge of recent DNN-based object detection is that it requires high computational costs. This requirement makes it challenging to deploy the DNN-based system on a vehicle for real-time inferencing. The low response time and high accuracy of automotive applications are critical factors when the system is deployed in real time. In this paper, the authors focus on deploying the computer-vision-based object detection system on the real-time service for automotive applications. First, five different vehicle detection systems are developed using transfer learning technology, which utilizes the pre-trained DNN model. The best performing DNN model showed improvements of 7.1% in Precision, 10.8% in Recall, and 8.93% in F1 score compared to the original YOLOv3 model. The developed DNN model was optimized by fusing layers horizontally and vertically to deploy it in the in-vehicle computing device. Finally, the optimized DNN model is deployed on the embedded in-vehicle computing device to run the program in real-time. Through optimization, the optimized DNN model can run 35.082 fps (frames per second) on the NVIDIA Jetson AGA, 19.385 times faster than the unoptimized DNN model. The experimental results demonstrate that the optimized transferred DNN model achieved higher accuracy and faster processing time for vehicle detection, which is vital for deploying the ADAS system. |
format | Online Article Text |
id | pubmed-10142959 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-101429592023-04-29 An Optimized DNN Model for Real-Time Inferencing on an Embedded Device Park, Jungme Aryal, Pawan Mandumula, Sai Rithvick Asolkar, Ritwik Prasad Sensors (Basel) Article For many automotive functionalities in Advanced Driver Assist Systems (ADAS) and Autonomous Driving (AD), target objects are detected using state-of-the-art Deep Neural Network (DNN) technologies. However, the main challenge of recent DNN-based object detection is that it requires high computational costs. This requirement makes it challenging to deploy the DNN-based system on a vehicle for real-time inferencing. The low response time and high accuracy of automotive applications are critical factors when the system is deployed in real time. In this paper, the authors focus on deploying the computer-vision-based object detection system on the real-time service for automotive applications. First, five different vehicle detection systems are developed using transfer learning technology, which utilizes the pre-trained DNN model. The best performing DNN model showed improvements of 7.1% in Precision, 10.8% in Recall, and 8.93% in F1 score compared to the original YOLOv3 model. The developed DNN model was optimized by fusing layers horizontally and vertically to deploy it in the in-vehicle computing device. Finally, the optimized DNN model is deployed on the embedded in-vehicle computing device to run the program in real-time. Through optimization, the optimized DNN model can run 35.082 fps (frames per second) on the NVIDIA Jetson AGA, 19.385 times faster than the unoptimized DNN model. The experimental results demonstrate that the optimized transferred DNN model achieved higher accuracy and faster processing time for vehicle detection, which is vital for deploying the ADAS system. MDPI 2023-04-14 /pmc/articles/PMC10142959/ /pubmed/37112333 http://dx.doi.org/10.3390/s23083992 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Park, Jungme Aryal, Pawan Mandumula, Sai Rithvick Asolkar, Ritwik Prasad An Optimized DNN Model for Real-Time Inferencing on an Embedded Device |
title | An Optimized DNN Model for Real-Time Inferencing on an Embedded Device |
title_full | An Optimized DNN Model for Real-Time Inferencing on an Embedded Device |
title_fullStr | An Optimized DNN Model for Real-Time Inferencing on an Embedded Device |
title_full_unstemmed | An Optimized DNN Model for Real-Time Inferencing on an Embedded Device |
title_short | An Optimized DNN Model for Real-Time Inferencing on an Embedded Device |
title_sort | optimized dnn model for real-time inferencing on an embedded device |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10142959/ https://www.ncbi.nlm.nih.gov/pubmed/37112333 http://dx.doi.org/10.3390/s23083992 |
work_keys_str_mv | AT parkjungme anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT aryalpawan anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT mandumulasairithvick anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT asolkarritwikprasad anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT parkjungme optimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT aryalpawan optimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT mandumulasairithvick optimizeddnnmodelforrealtimeinferencingonanembeddeddevice AT asolkarritwikprasad optimizeddnnmodelforrealtimeinferencingonanembeddeddevice |