Cargando…
An improved YOLO v4 used for grape detection in unstructured environment
Visual recognition is the most critical function of a harvesting robot, and the accuracy of the harvesting action is based on the performance of visual recognition. However, unstructured environment, such as severe occlusion, fruits overlap, illumination changes, complex backgrounds, and even heavy...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10374324/ https://www.ncbi.nlm.nih.gov/pubmed/37521937 http://dx.doi.org/10.3389/fpls.2023.1209910 |
_version_ | 1785078749889298432 |
---|---|
author | Guo, Canzhi Zheng, Shiwu Cheng, Guanggui Zhang, Yue Ding, Jianning |
author_facet | Guo, Canzhi Zheng, Shiwu Cheng, Guanggui Zhang, Yue Ding, Jianning |
author_sort | Guo, Canzhi |
collection | PubMed |
description | Visual recognition is the most critical function of a harvesting robot, and the accuracy of the harvesting action is based on the performance of visual recognition. However, unstructured environment, such as severe occlusion, fruits overlap, illumination changes, complex backgrounds, and even heavy fog weather, pose series of serious challenges to the detection accuracy of the recognition algorithm. Hence, this paper proposes an improved YOLO v4 model, called YOLO v4+, to cope with the challenges brought by unstructured environment. The output of each Resblock_body in the backbone is processed using a simple, parameterless attention mechanism for full dimensional refinement of extracted features. Further, in order to alleviate the problem of feature information loss, a multi scale feature fusion module with fusion weight and jump connection structure was pro-posed. In addition, the focal loss function is adopted and the hyperparameters α, γ are adjusted to 0.75 and 2. The experimental results show that the average precision of the YOLO v4+ model is 94.25% and the F1 score is 93%, which is 3.35% and 3% higher than the original YOLO v4 respectively. Compared with several state-of-the-art detection models, YOLO v4+ not only has the highest comprehensive ability, but also has better generalization ability. Selecting the corresponding augmentation method for specific working condition can greatly improve the model detection accuracy. Applying the proposed method to harvesting robots may enhance the applicability and robustness of the robotic system. |
format | Online Article Text |
id | pubmed-10374324 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-103743242023-07-28 An improved YOLO v4 used for grape detection in unstructured environment Guo, Canzhi Zheng, Shiwu Cheng, Guanggui Zhang, Yue Ding, Jianning Front Plant Sci Plant Science Visual recognition is the most critical function of a harvesting robot, and the accuracy of the harvesting action is based on the performance of visual recognition. However, unstructured environment, such as severe occlusion, fruits overlap, illumination changes, complex backgrounds, and even heavy fog weather, pose series of serious challenges to the detection accuracy of the recognition algorithm. Hence, this paper proposes an improved YOLO v4 model, called YOLO v4+, to cope with the challenges brought by unstructured environment. The output of each Resblock_body in the backbone is processed using a simple, parameterless attention mechanism for full dimensional refinement of extracted features. Further, in order to alleviate the problem of feature information loss, a multi scale feature fusion module with fusion weight and jump connection structure was pro-posed. In addition, the focal loss function is adopted and the hyperparameters α, γ are adjusted to 0.75 and 2. The experimental results show that the average precision of the YOLO v4+ model is 94.25% and the F1 score is 93%, which is 3.35% and 3% higher than the original YOLO v4 respectively. Compared with several state-of-the-art detection models, YOLO v4+ not only has the highest comprehensive ability, but also has better generalization ability. Selecting the corresponding augmentation method for specific working condition can greatly improve the model detection accuracy. Applying the proposed method to harvesting robots may enhance the applicability and robustness of the robotic system. Frontiers Media S.A. 2023-07-13 /pmc/articles/PMC10374324/ /pubmed/37521937 http://dx.doi.org/10.3389/fpls.2023.1209910 Text en Copyright © 2023 Guo, Zheng, Cheng, Zhang and Ding https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Plant Science Guo, Canzhi Zheng, Shiwu Cheng, Guanggui Zhang, Yue Ding, Jianning An improved YOLO v4 used for grape detection in unstructured environment |
title | An improved YOLO v4 used for grape detection in unstructured environment |
title_full | An improved YOLO v4 used for grape detection in unstructured environment |
title_fullStr | An improved YOLO v4 used for grape detection in unstructured environment |
title_full_unstemmed | An improved YOLO v4 used for grape detection in unstructured environment |
title_short | An improved YOLO v4 used for grape detection in unstructured environment |
title_sort | improved yolo v4 used for grape detection in unstructured environment |
topic | Plant Science |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10374324/ https://www.ncbi.nlm.nih.gov/pubmed/37521937 http://dx.doi.org/10.3389/fpls.2023.1209910 |
work_keys_str_mv | AT guocanzhi animprovedyolov4usedforgrapedetectioninunstructuredenvironment AT zhengshiwu animprovedyolov4usedforgrapedetectioninunstructuredenvironment AT chengguanggui animprovedyolov4usedforgrapedetectioninunstructuredenvironment AT zhangyue animprovedyolov4usedforgrapedetectioninunstructuredenvironment AT dingjianning animprovedyolov4usedforgrapedetectioninunstructuredenvironment AT guocanzhi improvedyolov4usedforgrapedetectioninunstructuredenvironment AT zhengshiwu improvedyolov4usedforgrapedetectioninunstructuredenvironment AT chengguanggui improvedyolov4usedforgrapedetectioninunstructuredenvironment AT zhangyue improvedyolov4usedforgrapedetectioninunstructuredenvironment AT dingjianning improvedyolov4usedforgrapedetectioninunstructuredenvironment |