Cargando…
Deep learning based object detection and surrounding environment description for visually impaired people
Object detection, one of the most significant contributions of computer vision and machine learning, plays an immense role in identifying and locating objects in an image or a video. We recognize distinct objects and precisely get their information through object detection, such as their size, shape...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10360957/ https://www.ncbi.nlm.nih.gov/pubmed/37484219 http://dx.doi.org/10.1016/j.heliyon.2023.e16924 |
_version_ | 1785076128256360448 |
---|---|
author | Islam, Raihan Bin Akhter, Samiha Iqbal, Faria Saif Ur Rahman, Md. Khan, Riasat |
author_facet | Islam, Raihan Bin Akhter, Samiha Iqbal, Faria Saif Ur Rahman, Md. Khan, Riasat |
author_sort | Islam, Raihan Bin |
collection | PubMed |
description | Object detection, one of the most significant contributions of computer vision and machine learning, plays an immense role in identifying and locating objects in an image or a video. We recognize distinct objects and precisely get their information through object detection, such as their size, shape, and location. This paper developed a low-cost assistive system of obstacle detection and the surrounding environment depiction to help blind people using deep learning techniques. TensorFlow object detection API and SSDLite MobileNetV2 have been used to create the proposed object detection model. The pre-trained SSDLite MobileNetV2 model is trained on the COCO dataset, with almost 328,000 images of 90 different objects. The gradient particle swarm optimization (PSO) technique has been used in this work to optimize the final layers and their corresponding hyperparameters of the MobileNetV2 model. Next, we used the Google text-to-speech module, PyAudio, playsound, and speech recognition to generate the audio feedback of the detected objects. A Raspberry Pi camera captures real-time video where real-time object detection is done frame by frame with Raspberry Pi 4B microcontroller. The proposed device is integrated into a head cap, which will help visually impaired people to detect obstacles in their path, as it is more efficient than a traditional white cane. Apart from this detection model, we trained a secondary computer vision model and named it the “ambiance mode.” In this mode, the last three convolutional layers of SSDLite MobileNetV2 are trained through transfer learning on a weather dataset. The dataset comprises around 500 images from four classes: cloudy, rainy, foggy, and sunrise. In this mode, the proposed system will narrate the surrounding scene elaborately, almost like a human describing a landscape or a beautiful sunset to a visually impaired person. The performance of the object detection and ambiance description modes are tested and evaluated in a desktop computer and Raspberry Pi embedded system. Detection accuracy and mean average precision, frame rate, confusion matrix, and ROC curve measure the model's accuracy on both setups. This low-cost proposed system is believed to help visually impaired people in their day-to-day life. |
format | Online Article Text |
id | pubmed-10360957 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-103609572023-07-22 Deep learning based object detection and surrounding environment description for visually impaired people Islam, Raihan Bin Akhter, Samiha Iqbal, Faria Saif Ur Rahman, Md. Khan, Riasat Heliyon Research Article Object detection, one of the most significant contributions of computer vision and machine learning, plays an immense role in identifying and locating objects in an image or a video. We recognize distinct objects and precisely get their information through object detection, such as their size, shape, and location. This paper developed a low-cost assistive system of obstacle detection and the surrounding environment depiction to help blind people using deep learning techniques. TensorFlow object detection API and SSDLite MobileNetV2 have been used to create the proposed object detection model. The pre-trained SSDLite MobileNetV2 model is trained on the COCO dataset, with almost 328,000 images of 90 different objects. The gradient particle swarm optimization (PSO) technique has been used in this work to optimize the final layers and their corresponding hyperparameters of the MobileNetV2 model. Next, we used the Google text-to-speech module, PyAudio, playsound, and speech recognition to generate the audio feedback of the detected objects. A Raspberry Pi camera captures real-time video where real-time object detection is done frame by frame with Raspberry Pi 4B microcontroller. The proposed device is integrated into a head cap, which will help visually impaired people to detect obstacles in their path, as it is more efficient than a traditional white cane. Apart from this detection model, we trained a secondary computer vision model and named it the “ambiance mode.” In this mode, the last three convolutional layers of SSDLite MobileNetV2 are trained through transfer learning on a weather dataset. The dataset comprises around 500 images from four classes: cloudy, rainy, foggy, and sunrise. In this mode, the proposed system will narrate the surrounding scene elaborately, almost like a human describing a landscape or a beautiful sunset to a visually impaired person. The performance of the object detection and ambiance description modes are tested and evaluated in a desktop computer and Raspberry Pi embedded system. Detection accuracy and mean average precision, frame rate, confusion matrix, and ROC curve measure the model's accuracy on both setups. This low-cost proposed system is believed to help visually impaired people in their day-to-day life. Elsevier 2023-06-07 /pmc/articles/PMC10360957/ /pubmed/37484219 http://dx.doi.org/10.1016/j.heliyon.2023.e16924 Text en © 2023 The Author(s) https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Research Article Islam, Raihan Bin Akhter, Samiha Iqbal, Faria Saif Ur Rahman, Md. Khan, Riasat Deep learning based object detection and surrounding environment description for visually impaired people |
title | Deep learning based object detection and surrounding environment description for visually impaired people |
title_full | Deep learning based object detection and surrounding environment description for visually impaired people |
title_fullStr | Deep learning based object detection and surrounding environment description for visually impaired people |
title_full_unstemmed | Deep learning based object detection and surrounding environment description for visually impaired people |
title_short | Deep learning based object detection and surrounding environment description for visually impaired people |
title_sort | deep learning based object detection and surrounding environment description for visually impaired people |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10360957/ https://www.ncbi.nlm.nih.gov/pubmed/37484219 http://dx.doi.org/10.1016/j.heliyon.2023.e16924 |
work_keys_str_mv | AT islamraihanbin deeplearningbasedobjectdetectionandsurroundingenvironmentdescriptionforvisuallyimpairedpeople AT akhtersamiha deeplearningbasedobjectdetectionandsurroundingenvironmentdescriptionforvisuallyimpairedpeople AT iqbalfaria deeplearningbasedobjectdetectionandsurroundingenvironmentdescriptionforvisuallyimpairedpeople AT saifurrahmanmd deeplearningbasedobjectdetectionandsurroundingenvironmentdescriptionforvisuallyimpairedpeople AT khanriasat deeplearningbasedobjectdetectionandsurroundingenvironmentdescriptionforvisuallyimpairedpeople |