Cargando…
Deep Active Learning via Open-Set Recognition
In many applications, data is easy to acquire but expensive and time-consuming to label, prominent examples include medical imaging and NLP. This disparity has only grown in recent years as our ability to collect data improves. Under these constraints, it makes sense to select only the most informat...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8859322/ https://www.ncbi.nlm.nih.gov/pubmed/35198969 http://dx.doi.org/10.3389/frai.2022.737363 |
_version_ | 1784654432161497088 |
---|---|
author | Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando |
author_facet | Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando |
author_sort | Mandivarapu, Jaya Krishna |
collection | PubMed |
description | In many applications, data is easy to acquire but expensive and time-consuming to label, prominent examples include medical imaging and NLP. This disparity has only grown in recent years as our ability to collect data improves. Under these constraints, it makes sense to select only the most informative instances from the unlabeled pool and request an oracle (e.g., a human expert) to provide labels for those samples. The goal of active learning is to infer the informativeness of unlabeled samples so as to minimize the number of requests to the oracle. Here, we formulate active learning as an open-set recognition problem. In this paradigm, only some of the inputs belong to known classes; the classifier must identify the rest as unknown. More specifically, we leverage variational neural networks (VNNs), which produce high-confidence (i.e., low-entropy) predictions only for inputs that closely resemble the training data. We use the inverse of this confidence measure to select the samples that the oracle should label. Intuitively, unlabeled samples that the VNN is uncertain about contain features that the network has not been exposed to; thus they are more informative for future training. We carried out an extensive evaluation of our novel, probabilistic formulation of active learning, achieving state-of-the-art results on MNIST, CIFAR-10, CIFAR-100, and FashionMNIST. Additionally, unlike current active learning methods, our algorithm can learn even in the presence of out-of-distribution outliers. As our experiments show, when the unlabeled pool consists of a mixture of samples from multiple datasets, our approach can automatically distinguish between samples from seen vs. unseen datasets. Overall, our results show that high-quality uncertainty measures are key for pool-based active learning. |
format | Online Article Text |
id | pubmed-8859322 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-88593222022-02-22 Deep Active Learning via Open-Set Recognition Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando Front Artif Intell Artificial Intelligence In many applications, data is easy to acquire but expensive and time-consuming to label, prominent examples include medical imaging and NLP. This disparity has only grown in recent years as our ability to collect data improves. Under these constraints, it makes sense to select only the most informative instances from the unlabeled pool and request an oracle (e.g., a human expert) to provide labels for those samples. The goal of active learning is to infer the informativeness of unlabeled samples so as to minimize the number of requests to the oracle. Here, we formulate active learning as an open-set recognition problem. In this paradigm, only some of the inputs belong to known classes; the classifier must identify the rest as unknown. More specifically, we leverage variational neural networks (VNNs), which produce high-confidence (i.e., low-entropy) predictions only for inputs that closely resemble the training data. We use the inverse of this confidence measure to select the samples that the oracle should label. Intuitively, unlabeled samples that the VNN is uncertain about contain features that the network has not been exposed to; thus they are more informative for future training. We carried out an extensive evaluation of our novel, probabilistic formulation of active learning, achieving state-of-the-art results on MNIST, CIFAR-10, CIFAR-100, and FashionMNIST. Additionally, unlike current active learning methods, our algorithm can learn even in the presence of out-of-distribution outliers. As our experiments show, when the unlabeled pool consists of a mixture of samples from multiple datasets, our approach can automatically distinguish between samples from seen vs. unseen datasets. Overall, our results show that high-quality uncertainty measures are key for pool-based active learning. Frontiers Media S.A. 2022-02-07 /pmc/articles/PMC8859322/ /pubmed/35198969 http://dx.doi.org/10.3389/frai.2022.737363 Text en Copyright © 2022 Mandivarapu, Camp and Estrada. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando Deep Active Learning via Open-Set Recognition |
title | Deep Active Learning via Open-Set Recognition |
title_full | Deep Active Learning via Open-Set Recognition |
title_fullStr | Deep Active Learning via Open-Set Recognition |
title_full_unstemmed | Deep Active Learning via Open-Set Recognition |
title_short | Deep Active Learning via Open-Set Recognition |
title_sort | deep active learning via open-set recognition |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8859322/ https://www.ncbi.nlm.nih.gov/pubmed/35198969 http://dx.doi.org/10.3389/frai.2022.737363 |
work_keys_str_mv | AT mandivarapujayakrishna deepactivelearningviaopensetrecognition AT campblake deepactivelearningviaopensetrecognition AT estradarolando deepactivelearningviaopensetrecognition |