Cargando…

Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion

Heavily occluded objects are more difficult for classification algorithms to identify correctly than unoccluded objects. This effect is rare and thus hard to measure with datasets like ImageNet and PASCAL VOC, however, owing to biases in human-generated image pose selection. We introduce a dataset t...

Descripción completa

Detalles Bibliográficos
Autores principales: Chandler, Benjamin, Mingolla, Ennio
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi Publishing Corporation 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4908250/
https://www.ncbi.nlm.nih.gov/pubmed/27340396
http://dx.doi.org/10.1155/2016/6425257
_version_ 1782437646704836608
author Chandler, Benjamin
Mingolla, Ennio
author_facet Chandler, Benjamin
Mingolla, Ennio
author_sort Chandler, Benjamin
collection PubMed
description Heavily occluded objects are more difficult for classification algorithms to identify correctly than unoccluded objects. This effect is rare and thus hard to measure with datasets like ImageNet and PASCAL VOC, however, owing to biases in human-generated image pose selection. We introduce a dataset that emphasizes occlusion and additions to a standard convolutional neural network aimed at increasing invariance to occlusion. An unmodified convolutional neural network trained and tested on the new dataset rapidly degrades to chance-level accuracy as occlusion increases. Training with occluded data slows this decline but still yields poor performance with high occlusion. Integrating novel preprocessing stages to segment the input and inpaint occlusions is an effective mitigation. A convolutional network so modified is nearly as effective with more than 81% of pixels occluded as it is with no occlusion. Such a network is also more accurate on unoccluded images than an otherwise identical network that has been trained with only unoccluded images. These results depend on successful segmentation. The occlusions in our dataset are deliberately easy to segment from the figure and background. Achieving similar results on a more challenging dataset would require finding a method to split figure, background, and occluding pixels in the input.
format Online
Article
Text
id pubmed-4908250
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher Hindawi Publishing Corporation
record_format MEDLINE/PubMed
spelling pubmed-49082502016-06-23 Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion Chandler, Benjamin Mingolla, Ennio Comput Intell Neurosci Research Article Heavily occluded objects are more difficult for classification algorithms to identify correctly than unoccluded objects. This effect is rare and thus hard to measure with datasets like ImageNet and PASCAL VOC, however, owing to biases in human-generated image pose selection. We introduce a dataset that emphasizes occlusion and additions to a standard convolutional neural network aimed at increasing invariance to occlusion. An unmodified convolutional neural network trained and tested on the new dataset rapidly degrades to chance-level accuracy as occlusion increases. Training with occluded data slows this decline but still yields poor performance with high occlusion. Integrating novel preprocessing stages to segment the input and inpaint occlusions is an effective mitigation. A convolutional network so modified is nearly as effective with more than 81% of pixels occluded as it is with no occlusion. Such a network is also more accurate on unoccluded images than an otherwise identical network that has been trained with only unoccluded images. These results depend on successful segmentation. The occlusions in our dataset are deliberately easy to segment from the figure and background. Achieving similar results on a more challenging dataset would require finding a method to split figure, background, and occluding pixels in the input. Hindawi Publishing Corporation 2016 2016-06-01 /pmc/articles/PMC4908250/ /pubmed/27340396 http://dx.doi.org/10.1155/2016/6425257 Text en Copyright © 2016 B. Chandler and E. Mingolla. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Chandler, Benjamin
Mingolla, Ennio
Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion
title Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion
title_full Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion
title_fullStr Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion
title_full_unstemmed Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion
title_short Mitigation of Effects of Occlusion on Object Recognition with Deep Neural Networks through Low-Level Image Completion
title_sort mitigation of effects of occlusion on object recognition with deep neural networks through low-level image completion
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4908250/
https://www.ncbi.nlm.nih.gov/pubmed/27340396
http://dx.doi.org/10.1155/2016/6425257
work_keys_str_mv AT chandlerbenjamin mitigationofeffectsofocclusiononobjectrecognitionwithdeepneuralnetworksthroughlowlevelimagecompletion
AT mingollaennio mitigationofeffectsofocclusiononobjectrecognitionwithdeepneuralnetworksthroughlowlevelimagecompletion