Cargando…

An improved YOLO v4 used for grape detection in unstructured environment

Visual recognition is the most critical function of a harvesting robot, and the accuracy of the harvesting action is based on the performance of visual recognition. However, unstructured environment, such as severe occlusion, fruits overlap, illumination changes, complex backgrounds, and even heavy...

Descripción completa

Detalles Bibliográficos
Autores principales: Guo, Canzhi, Zheng, Shiwu, Cheng, Guanggui, Zhang, Yue, Ding, Jianning
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10374324/
https://www.ncbi.nlm.nih.gov/pubmed/37521937
http://dx.doi.org/10.3389/fpls.2023.1209910
_version_ 1785078749889298432
author Guo, Canzhi
Zheng, Shiwu
Cheng, Guanggui
Zhang, Yue
Ding, Jianning
author_facet Guo, Canzhi
Zheng, Shiwu
Cheng, Guanggui
Zhang, Yue
Ding, Jianning
author_sort Guo, Canzhi
collection PubMed
description Visual recognition is the most critical function of a harvesting robot, and the accuracy of the harvesting action is based on the performance of visual recognition. However, unstructured environment, such as severe occlusion, fruits overlap, illumination changes, complex backgrounds, and even heavy fog weather, pose series of serious challenges to the detection accuracy of the recognition algorithm. Hence, this paper proposes an improved YOLO v4 model, called YOLO v4+, to cope with the challenges brought by unstructured environment. The output of each Resblock_body in the backbone is processed using a simple, parameterless attention mechanism for full dimensional refinement of extracted features. Further, in order to alleviate the problem of feature information loss, a multi scale feature fusion module with fusion weight and jump connection structure was pro-posed. In addition, the focal loss function is adopted and the hyperparameters α, γ are adjusted to 0.75 and 2. The experimental results show that the average precision of the YOLO v4+ model is 94.25% and the F1 score is 93%, which is 3.35% and 3% higher than the original YOLO v4 respectively. Compared with several state-of-the-art detection models, YOLO v4+ not only has the highest comprehensive ability, but also has better generalization ability. Selecting the corresponding augmentation method for specific working condition can greatly improve the model detection accuracy. Applying the proposed method to harvesting robots may enhance the applicability and robustness of the robotic system.
format Online
Article
Text
id pubmed-10374324
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-103743242023-07-28 An improved YOLO v4 used for grape detection in unstructured environment Guo, Canzhi Zheng, Shiwu Cheng, Guanggui Zhang, Yue Ding, Jianning Front Plant Sci Plant Science Visual recognition is the most critical function of a harvesting robot, and the accuracy of the harvesting action is based on the performance of visual recognition. However, unstructured environment, such as severe occlusion, fruits overlap, illumination changes, complex backgrounds, and even heavy fog weather, pose series of serious challenges to the detection accuracy of the recognition algorithm. Hence, this paper proposes an improved YOLO v4 model, called YOLO v4+, to cope with the challenges brought by unstructured environment. The output of each Resblock_body in the backbone is processed using a simple, parameterless attention mechanism for full dimensional refinement of extracted features. Further, in order to alleviate the problem of feature information loss, a multi scale feature fusion module with fusion weight and jump connection structure was pro-posed. In addition, the focal loss function is adopted and the hyperparameters α, γ are adjusted to 0.75 and 2. The experimental results show that the average precision of the YOLO v4+ model is 94.25% and the F1 score is 93%, which is 3.35% and 3% higher than the original YOLO v4 respectively. Compared with several state-of-the-art detection models, YOLO v4+ not only has the highest comprehensive ability, but also has better generalization ability. Selecting the corresponding augmentation method for specific working condition can greatly improve the model detection accuracy. Applying the proposed method to harvesting robots may enhance the applicability and robustness of the robotic system. Frontiers Media S.A. 2023-07-13 /pmc/articles/PMC10374324/ /pubmed/37521937 http://dx.doi.org/10.3389/fpls.2023.1209910 Text en Copyright © 2023 Guo, Zheng, Cheng, Zhang and Ding https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Plant Science
Guo, Canzhi
Zheng, Shiwu
Cheng, Guanggui
Zhang, Yue
Ding, Jianning
An improved YOLO v4 used for grape detection in unstructured environment
title An improved YOLO v4 used for grape detection in unstructured environment
title_full An improved YOLO v4 used for grape detection in unstructured environment
title_fullStr An improved YOLO v4 used for grape detection in unstructured environment
title_full_unstemmed An improved YOLO v4 used for grape detection in unstructured environment
title_short An improved YOLO v4 used for grape detection in unstructured environment
title_sort improved yolo v4 used for grape detection in unstructured environment
topic Plant Science
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10374324/
https://www.ncbi.nlm.nih.gov/pubmed/37521937
http://dx.doi.org/10.3389/fpls.2023.1209910
work_keys_str_mv AT guocanzhi animprovedyolov4usedforgrapedetectioninunstructuredenvironment
AT zhengshiwu animprovedyolov4usedforgrapedetectioninunstructuredenvironment
AT chengguanggui animprovedyolov4usedforgrapedetectioninunstructuredenvironment
AT zhangyue animprovedyolov4usedforgrapedetectioninunstructuredenvironment
AT dingjianning animprovedyolov4usedforgrapedetectioninunstructuredenvironment
AT guocanzhi improvedyolov4usedforgrapedetectioninunstructuredenvironment
AT zhengshiwu improvedyolov4usedforgrapedetectioninunstructuredenvironment
AT chengguanggui improvedyolov4usedforgrapedetectioninunstructuredenvironment
AT zhangyue improvedyolov4usedforgrapedetectioninunstructuredenvironment
AT dingjianning improvedyolov4usedforgrapedetectioninunstructuredenvironment