Cargando…

An Optimized DNN Model for Real-Time Inferencing on an Embedded Device

For many automotive functionalities in Advanced Driver Assist Systems (ADAS) and Autonomous Driving (AD), target objects are detected using state-of-the-art Deep Neural Network (DNN) technologies. However, the main challenge of recent DNN-based object detection is that it requires high computational...

Descripción completa

Detalles Bibliográficos
Autores principales: Park, Jungme, Aryal, Pawan, Mandumula, Sai Rithvick, Asolkar, Ritwik Prasad
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10142959/
https://www.ncbi.nlm.nih.gov/pubmed/37112333
http://dx.doi.org/10.3390/s23083992
_version_ 1785033737143058432
author Park, Jungme
Aryal, Pawan
Mandumula, Sai Rithvick
Asolkar, Ritwik Prasad
author_facet Park, Jungme
Aryal, Pawan
Mandumula, Sai Rithvick
Asolkar, Ritwik Prasad
author_sort Park, Jungme
collection PubMed
description For many automotive functionalities in Advanced Driver Assist Systems (ADAS) and Autonomous Driving (AD), target objects are detected using state-of-the-art Deep Neural Network (DNN) technologies. However, the main challenge of recent DNN-based object detection is that it requires high computational costs. This requirement makes it challenging to deploy the DNN-based system on a vehicle for real-time inferencing. The low response time and high accuracy of automotive applications are critical factors when the system is deployed in real time. In this paper, the authors focus on deploying the computer-vision-based object detection system on the real-time service for automotive applications. First, five different vehicle detection systems are developed using transfer learning technology, which utilizes the pre-trained DNN model. The best performing DNN model showed improvements of 7.1% in Precision, 10.8% in Recall, and 8.93% in F1 score compared to the original YOLOv3 model. The developed DNN model was optimized by fusing layers horizontally and vertically to deploy it in the in-vehicle computing device. Finally, the optimized DNN model is deployed on the embedded in-vehicle computing device to run the program in real-time. Through optimization, the optimized DNN model can run 35.082 fps (frames per second) on the NVIDIA Jetson AGA, 19.385 times faster than the unoptimized DNN model. The experimental results demonstrate that the optimized transferred DNN model achieved higher accuracy and faster processing time for vehicle detection, which is vital for deploying the ADAS system.
format Online
Article
Text
id pubmed-10142959
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-101429592023-04-29 An Optimized DNN Model for Real-Time Inferencing on an Embedded Device Park, Jungme Aryal, Pawan Mandumula, Sai Rithvick Asolkar, Ritwik Prasad Sensors (Basel) Article For many automotive functionalities in Advanced Driver Assist Systems (ADAS) and Autonomous Driving (AD), target objects are detected using state-of-the-art Deep Neural Network (DNN) technologies. However, the main challenge of recent DNN-based object detection is that it requires high computational costs. This requirement makes it challenging to deploy the DNN-based system on a vehicle for real-time inferencing. The low response time and high accuracy of automotive applications are critical factors when the system is deployed in real time. In this paper, the authors focus on deploying the computer-vision-based object detection system on the real-time service for automotive applications. First, five different vehicle detection systems are developed using transfer learning technology, which utilizes the pre-trained DNN model. The best performing DNN model showed improvements of 7.1% in Precision, 10.8% in Recall, and 8.93% in F1 score compared to the original YOLOv3 model. The developed DNN model was optimized by fusing layers horizontally and vertically to deploy it in the in-vehicle computing device. Finally, the optimized DNN model is deployed on the embedded in-vehicle computing device to run the program in real-time. Through optimization, the optimized DNN model can run 35.082 fps (frames per second) on the NVIDIA Jetson AGA, 19.385 times faster than the unoptimized DNN model. The experimental results demonstrate that the optimized transferred DNN model achieved higher accuracy and faster processing time for vehicle detection, which is vital for deploying the ADAS system. MDPI 2023-04-14 /pmc/articles/PMC10142959/ /pubmed/37112333 http://dx.doi.org/10.3390/s23083992 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Park, Jungme
Aryal, Pawan
Mandumula, Sai Rithvick
Asolkar, Ritwik Prasad
An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
title An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
title_full An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
title_fullStr An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
title_full_unstemmed An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
title_short An Optimized DNN Model for Real-Time Inferencing on an Embedded Device
title_sort optimized dnn model for real-time inferencing on an embedded device
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10142959/
https://www.ncbi.nlm.nih.gov/pubmed/37112333
http://dx.doi.org/10.3390/s23083992
work_keys_str_mv AT parkjungme anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT aryalpawan anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT mandumulasairithvick anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT asolkarritwikprasad anoptimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT parkjungme optimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT aryalpawan optimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT mandumulasairithvick optimizeddnnmodelforrealtimeinferencingonanembeddeddevice
AT asolkarritwikprasad optimizeddnnmodelforrealtimeinferencingonanembeddeddevice