Cargando…

DeepFruits: A Fruit Detection System Using Deep Neural Networks

This paper presents a novel approach to fruit detection using deep convolutional neural networks. The aim is to build an accurate, fast and reliable fruit detection system, which is a vital element of an autonomous agricultural robotic platform; it is a key element for fruit yield estimation and aut...

Descripción completa

Detalles Bibliográficos
Autores principales: Sa, Inkyu, Ge, Zongyuan, Dayoub, Feras, Upcroft, Ben, Perez, Tristan, McCool, Chris
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5017387/
https://www.ncbi.nlm.nih.gov/pubmed/27527168
http://dx.doi.org/10.3390/s16081222
_version_ 1782452736451674112
author Sa, Inkyu
Ge, Zongyuan
Dayoub, Feras
Upcroft, Ben
Perez, Tristan
McCool, Chris
author_facet Sa, Inkyu
Ge, Zongyuan
Dayoub, Feras
Upcroft, Ben
Perez, Tristan
McCool, Chris
author_sort Sa, Inkyu
collection PubMed
description This paper presents a novel approach to fruit detection using deep convolutional neural networks. The aim is to build an accurate, fast and reliable fruit detection system, which is a vital element of an autonomous agricultural robotic platform; it is a key element for fruit yield estimation and automated harvesting. Recent work in deep neural networks has led to the development of a state-of-the-art object detector termed Faster Region-based CNN (Faster R-CNN). We adapt this model, through transfer learning, for the task of fruit detection using imagery obtained from two modalities: colour (RGB) and Near-Infrared (NIR). Early and late fusion methods are explored for combining the multi-modal (RGB and NIR) information. This leads to a novel multi-modal Faster R-CNN model, which achieves state-of-the-art results compared to prior work with the F1 score, which takes into account both precision and recall performances improving from [Formula: see text] to [Formula: see text] for the detection of sweet pepper. In addition to improved accuracy, this approach is also much quicker to deploy for new fruits, as it requires bounding box annotation rather than pixel-level annotation (annotating bounding boxes is approximately an order of magnitude quicker to perform). The model is retrained to perform the detection of seven fruits, with the entire process taking four hours to annotate and train the new model per fruit.
format Online
Article
Text
id pubmed-5017387
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-50173872016-09-22 DeepFruits: A Fruit Detection System Using Deep Neural Networks Sa, Inkyu Ge, Zongyuan Dayoub, Feras Upcroft, Ben Perez, Tristan McCool, Chris Sensors (Basel) Article This paper presents a novel approach to fruit detection using deep convolutional neural networks. The aim is to build an accurate, fast and reliable fruit detection system, which is a vital element of an autonomous agricultural robotic platform; it is a key element for fruit yield estimation and automated harvesting. Recent work in deep neural networks has led to the development of a state-of-the-art object detector termed Faster Region-based CNN (Faster R-CNN). We adapt this model, through transfer learning, for the task of fruit detection using imagery obtained from two modalities: colour (RGB) and Near-Infrared (NIR). Early and late fusion methods are explored for combining the multi-modal (RGB and NIR) information. This leads to a novel multi-modal Faster R-CNN model, which achieves state-of-the-art results compared to prior work with the F1 score, which takes into account both precision and recall performances improving from [Formula: see text] to [Formula: see text] for the detection of sweet pepper. In addition to improved accuracy, this approach is also much quicker to deploy for new fruits, as it requires bounding box annotation rather than pixel-level annotation (annotating bounding boxes is approximately an order of magnitude quicker to perform). The model is retrained to perform the detection of seven fruits, with the entire process taking four hours to annotate and train the new model per fruit. MDPI 2016-08-03 /pmc/articles/PMC5017387/ /pubmed/27527168 http://dx.doi.org/10.3390/s16081222 Text en © 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sa, Inkyu
Ge, Zongyuan
Dayoub, Feras
Upcroft, Ben
Perez, Tristan
McCool, Chris
DeepFruits: A Fruit Detection System Using Deep Neural Networks
title DeepFruits: A Fruit Detection System Using Deep Neural Networks
title_full DeepFruits: A Fruit Detection System Using Deep Neural Networks
title_fullStr DeepFruits: A Fruit Detection System Using Deep Neural Networks
title_full_unstemmed DeepFruits: A Fruit Detection System Using Deep Neural Networks
title_short DeepFruits: A Fruit Detection System Using Deep Neural Networks
title_sort deepfruits: a fruit detection system using deep neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5017387/
https://www.ncbi.nlm.nih.gov/pubmed/27527168
http://dx.doi.org/10.3390/s16081222
work_keys_str_mv AT sainkyu deepfruitsafruitdetectionsystemusingdeepneuralnetworks
AT gezongyuan deepfruitsafruitdetectionsystemusingdeepneuralnetworks
AT dayoubferas deepfruitsafruitdetectionsystemusingdeepneuralnetworks
AT upcroftben deepfruitsafruitdetectionsystemusingdeepneuralnetworks
AT pereztristan deepfruitsafruitdetectionsystemusingdeepneuralnetworks
AT mccoolchris deepfruitsafruitdetectionsystemusingdeepneuralnetworks