Cargando…

Deep Active Learning via Open-Set Recognition

In many applications, data is easy to acquire but expensive and time-consuming to label, prominent examples include medical imaging and NLP. This disparity has only grown in recent years as our ability to collect data improves. Under these constraints, it makes sense to select only the most informat...

Descripción completa

Detalles Bibliográficos
Autores principales: Mandivarapu, Jaya Krishna, Camp, Blake, Estrada, Rolando
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8859322/
https://www.ncbi.nlm.nih.gov/pubmed/35198969
http://dx.doi.org/10.3389/frai.2022.737363
_version_ 1784654432161497088
author Mandivarapu, Jaya Krishna
Camp, Blake
Estrada, Rolando
author_facet Mandivarapu, Jaya Krishna
Camp, Blake
Estrada, Rolando
author_sort Mandivarapu, Jaya Krishna
collection PubMed
description In many applications, data is easy to acquire but expensive and time-consuming to label, prominent examples include medical imaging and NLP. This disparity has only grown in recent years as our ability to collect data improves. Under these constraints, it makes sense to select only the most informative instances from the unlabeled pool and request an oracle (e.g., a human expert) to provide labels for those samples. The goal of active learning is to infer the informativeness of unlabeled samples so as to minimize the number of requests to the oracle. Here, we formulate active learning as an open-set recognition problem. In this paradigm, only some of the inputs belong to known classes; the classifier must identify the rest as unknown. More specifically, we leverage variational neural networks (VNNs), which produce high-confidence (i.e., low-entropy) predictions only for inputs that closely resemble the training data. We use the inverse of this confidence measure to select the samples that the oracle should label. Intuitively, unlabeled samples that the VNN is uncertain about contain features that the network has not been exposed to; thus they are more informative for future training. We carried out an extensive evaluation of our novel, probabilistic formulation of active learning, achieving state-of-the-art results on MNIST, CIFAR-10, CIFAR-100, and FashionMNIST. Additionally, unlike current active learning methods, our algorithm can learn even in the presence of out-of-distribution outliers. As our experiments show, when the unlabeled pool consists of a mixture of samples from multiple datasets, our approach can automatically distinguish between samples from seen vs. unseen datasets. Overall, our results show that high-quality uncertainty measures are key for pool-based active learning.
format Online
Article
Text
id pubmed-8859322
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-88593222022-02-22 Deep Active Learning via Open-Set Recognition Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando Front Artif Intell Artificial Intelligence In many applications, data is easy to acquire but expensive and time-consuming to label, prominent examples include medical imaging and NLP. This disparity has only grown in recent years as our ability to collect data improves. Under these constraints, it makes sense to select only the most informative instances from the unlabeled pool and request an oracle (e.g., a human expert) to provide labels for those samples. The goal of active learning is to infer the informativeness of unlabeled samples so as to minimize the number of requests to the oracle. Here, we formulate active learning as an open-set recognition problem. In this paradigm, only some of the inputs belong to known classes; the classifier must identify the rest as unknown. More specifically, we leverage variational neural networks (VNNs), which produce high-confidence (i.e., low-entropy) predictions only for inputs that closely resemble the training data. We use the inverse of this confidence measure to select the samples that the oracle should label. Intuitively, unlabeled samples that the VNN is uncertain about contain features that the network has not been exposed to; thus they are more informative for future training. We carried out an extensive evaluation of our novel, probabilistic formulation of active learning, achieving state-of-the-art results on MNIST, CIFAR-10, CIFAR-100, and FashionMNIST. Additionally, unlike current active learning methods, our algorithm can learn even in the presence of out-of-distribution outliers. As our experiments show, when the unlabeled pool consists of a mixture of samples from multiple datasets, our approach can automatically distinguish between samples from seen vs. unseen datasets. Overall, our results show that high-quality uncertainty measures are key for pool-based active learning. Frontiers Media S.A. 2022-02-07 /pmc/articles/PMC8859322/ /pubmed/35198969 http://dx.doi.org/10.3389/frai.2022.737363 Text en Copyright © 2022 Mandivarapu, Camp and Estrada. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Mandivarapu, Jaya Krishna
Camp, Blake
Estrada, Rolando
Deep Active Learning via Open-Set Recognition
title Deep Active Learning via Open-Set Recognition
title_full Deep Active Learning via Open-Set Recognition
title_fullStr Deep Active Learning via Open-Set Recognition
title_full_unstemmed Deep Active Learning via Open-Set Recognition
title_short Deep Active Learning via Open-Set Recognition
title_sort deep active learning via open-set recognition
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8859322/
https://www.ncbi.nlm.nih.gov/pubmed/35198969
http://dx.doi.org/10.3389/frai.2022.737363
work_keys_str_mv AT mandivarapujayakrishna deepactivelearningviaopensetrecognition
AT campblake deepactivelearningviaopensetrecognition
AT estradarolando deepactivelearningviaopensetrecognition