Cargando…

Non-Deep Active Learning for Deep Neural Networks

One way to improve annotation efficiency is active learning. The goal of active learning is to select images from many unlabeled images, where labeling will improve the accuracy of the machine learning model the most. To select the most informative unlabeled images, conventional methods use deep neu...

Descripción completa

Detalles Bibliográficos
Autores principales: Kawano, Yasufumi, Nota, Yoshiki, Mochizuki, Rinpei, Aoki, Yoshimitsu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9319968/
https://www.ncbi.nlm.nih.gov/pubmed/35890924
http://dx.doi.org/10.3390/s22145244
_version_ 1784755679305662464
author Kawano, Yasufumi
Nota, Yoshiki
Mochizuki, Rinpei
Aoki, Yoshimitsu
author_facet Kawano, Yasufumi
Nota, Yoshiki
Mochizuki, Rinpei
Aoki, Yoshimitsu
author_sort Kawano, Yasufumi
collection PubMed
description One way to improve annotation efficiency is active learning. The goal of active learning is to select images from many unlabeled images, where labeling will improve the accuracy of the machine learning model the most. To select the most informative unlabeled images, conventional methods use deep neural networks with a large number of computation nodes and long computation time, but we propose a non-deep neural network method that does not require any additional training for unlabeled image selection. The proposed method trains a task model on labeled images, and then the model predicts unlabeled images. Based on this prediction, an uncertainty indicator is generated for each unlabeled image. Images with a high uncertainty index are considered to have a high information content, and are selected for annotation. Our proposed method is based on a very simple and powerful idea: select samples near the decision boundary of the model. Experimental results on multiple datasets show that the proposed method achieves higher accuracy than conventional active learning methods on multiple tasks and up to 14 times faster execution time from [Formula: see text] s to [Formula: see text] s. The proposed method outperforms the current SoTA method by 1% accuracy on CIFAR-10.
format Online
Article
Text
id pubmed-9319968
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-93199682022-07-27 Non-Deep Active Learning for Deep Neural Networks Kawano, Yasufumi Nota, Yoshiki Mochizuki, Rinpei Aoki, Yoshimitsu Sensors (Basel) Article One way to improve annotation efficiency is active learning. The goal of active learning is to select images from many unlabeled images, where labeling will improve the accuracy of the machine learning model the most. To select the most informative unlabeled images, conventional methods use deep neural networks with a large number of computation nodes and long computation time, but we propose a non-deep neural network method that does not require any additional training for unlabeled image selection. The proposed method trains a task model on labeled images, and then the model predicts unlabeled images. Based on this prediction, an uncertainty indicator is generated for each unlabeled image. Images with a high uncertainty index are considered to have a high information content, and are selected for annotation. Our proposed method is based on a very simple and powerful idea: select samples near the decision boundary of the model. Experimental results on multiple datasets show that the proposed method achieves higher accuracy than conventional active learning methods on multiple tasks and up to 14 times faster execution time from [Formula: see text] s to [Formula: see text] s. The proposed method outperforms the current SoTA method by 1% accuracy on CIFAR-10. MDPI 2022-07-13 /pmc/articles/PMC9319968/ /pubmed/35890924 http://dx.doi.org/10.3390/s22145244 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kawano, Yasufumi
Nota, Yoshiki
Mochizuki, Rinpei
Aoki, Yoshimitsu
Non-Deep Active Learning for Deep Neural Networks
title Non-Deep Active Learning for Deep Neural Networks
title_full Non-Deep Active Learning for Deep Neural Networks
title_fullStr Non-Deep Active Learning for Deep Neural Networks
title_full_unstemmed Non-Deep Active Learning for Deep Neural Networks
title_short Non-Deep Active Learning for Deep Neural Networks
title_sort non-deep active learning for deep neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9319968/
https://www.ncbi.nlm.nih.gov/pubmed/35890924
http://dx.doi.org/10.3390/s22145244
work_keys_str_mv AT kawanoyasufumi nondeepactivelearningfordeepneuralnetworks
AT notayoshiki nondeepactivelearningfordeepneuralnetworks
AT mochizukirinpei nondeepactivelearningfordeepneuralnetworks
AT aokiyoshimitsu nondeepactivelearningfordeepneuralnetworks