Cargando…

Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows

The relative position of the orchard robot to the rows of fruit trees is an important parameter for achieving autonomous navigation. The current methods for estimating the position parameters between rows of orchard robots obtain low parameter accuracy. To address this problem, this paper proposes a...

Descripción completa

Detalles Bibliográficos
Autores principales: Gu, Baoxing, Liu, Qin, Gao, Yi, Tian, Guangzhao, Zhang, Baohua, Wang, Haiqing, Li, He
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10650010/
https://www.ncbi.nlm.nih.gov/pubmed/37960506
http://dx.doi.org/10.3390/s23218807
_version_ 1785135681296662528
author Gu, Baoxing
Liu, Qin
Gao, Yi
Tian, Guangzhao
Zhang, Baohua
Wang, Haiqing
Li, He
author_facet Gu, Baoxing
Liu, Qin
Gao, Yi
Tian, Guangzhao
Zhang, Baohua
Wang, Haiqing
Li, He
author_sort Gu, Baoxing
collection PubMed
description The relative position of the orchard robot to the rows of fruit trees is an important parameter for achieving autonomous navigation. The current methods for estimating the position parameters between rows of orchard robots obtain low parameter accuracy. To address this problem, this paper proposes a machine vision-based method for detecting the relative position of orchard robots and fruit tree rows. First, the fruit tree trunk is identified based on the improved YOLOv4 model; second, the camera coordinates of the tree trunk are calculated using the principle of binocular camera triangulation, and the ground projection coordinates of the tree trunk are obtained through coordinate conversion; finally, the midpoints of the projection coordinates of different sides are combined, the navigation path is obtained by linear fitting with the least squares method, and the position parameters of the orchard robot are obtained through calculation. The experimental results show that the average accuracy and average recall rate of the improved YOLOv4 model for fruit tree trunk detection are 5.92% and 7.91% higher, respectively, than those of the original YOLOv4 model. The average errors of heading angle and lateral deviation estimates obtained based on the method in this paper are 0.57° and 0.02 m. The method can accurately calculate heading angle and lateral deviation values at different positions between rows and provide a reference for the autonomous visual navigation of orchard robots.
format Online
Article
Text
id pubmed-10650010
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-106500102023-10-29 Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows Gu, Baoxing Liu, Qin Gao, Yi Tian, Guangzhao Zhang, Baohua Wang, Haiqing Li, He Sensors (Basel) Article The relative position of the orchard robot to the rows of fruit trees is an important parameter for achieving autonomous navigation. The current methods for estimating the position parameters between rows of orchard robots obtain low parameter accuracy. To address this problem, this paper proposes a machine vision-based method for detecting the relative position of orchard robots and fruit tree rows. First, the fruit tree trunk is identified based on the improved YOLOv4 model; second, the camera coordinates of the tree trunk are calculated using the principle of binocular camera triangulation, and the ground projection coordinates of the tree trunk are obtained through coordinate conversion; finally, the midpoints of the projection coordinates of different sides are combined, the navigation path is obtained by linear fitting with the least squares method, and the position parameters of the orchard robot are obtained through calculation. The experimental results show that the average accuracy and average recall rate of the improved YOLOv4 model for fruit tree trunk detection are 5.92% and 7.91% higher, respectively, than those of the original YOLOv4 model. The average errors of heading angle and lateral deviation estimates obtained based on the method in this paper are 0.57° and 0.02 m. The method can accurately calculate heading angle and lateral deviation values at different positions between rows and provide a reference for the autonomous visual navigation of orchard robots. MDPI 2023-10-29 /pmc/articles/PMC10650010/ /pubmed/37960506 http://dx.doi.org/10.3390/s23218807 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Gu, Baoxing
Liu, Qin
Gao, Yi
Tian, Guangzhao
Zhang, Baohua
Wang, Haiqing
Li, He
Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows
title Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows
title_full Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows
title_fullStr Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows
title_full_unstemmed Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows
title_short Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows
title_sort research on the relative position detection method between orchard robots and fruit tree rows
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10650010/
https://www.ncbi.nlm.nih.gov/pubmed/37960506
http://dx.doi.org/10.3390/s23218807
work_keys_str_mv AT gubaoxing researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows
AT liuqin researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows
AT gaoyi researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows
AT tianguangzhao researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows
AT zhangbaohua researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows
AT wanghaiqing researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows
AT lihe researchontherelativepositiondetectionmethodbetweenorchardrobotsandfruittreerows