Cargando…

Evaluation of Uncertainty Quantification in Deep Learning

Artificial intelligence (AI) is nowadays included into an increasing number of critical systems. Inclusion of AI in such systems may, however, pose a risk, since it is, still, infeasible to build AI systems that know how to function well in situations that differ greatly from what the AI has seen be...

Descripción completa

Detalles Bibliográficos
Autores principales: Ståhl, Niclas, Falkman, Göran, Karlsson, Alexander, Mathiason, Gunnar
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7274324/
http://dx.doi.org/10.1007/978-3-030-50146-4_41
_version_ 1783542556547612672
author Ståhl, Niclas
Falkman, Göran
Karlsson, Alexander
Mathiason, Gunnar
author_facet Ståhl, Niclas
Falkman, Göran
Karlsson, Alexander
Mathiason, Gunnar
author_sort Ståhl, Niclas
collection PubMed
description Artificial intelligence (AI) is nowadays included into an increasing number of critical systems. Inclusion of AI in such systems may, however, pose a risk, since it is, still, infeasible to build AI systems that know how to function well in situations that differ greatly from what the AI has seen before. Therefore, it is crucial that future AI systems have the ability to not only function well in known domains, but also understand and show when they are uncertain when facing something unknown. In this paper, we evaluate four different methods that have been proposed to correctly quantifying uncertainty when the AI model is faced with new samples. We investigate the behaviour of these models when they are applied to samples far from what these models have seen before, and if they correctly attribute those samples with high uncertainty. We also examine if incorrectly classified samples are attributed with an higher uncertainty than correctly classified samples. The major finding from this simple experiment is, surprisingly, that the evaluated methods capture the uncertainty differently and the correlation between the quantified uncertainty of the models is low. This inconsistency is something that needs to be further understood and solved before AI can be used in critical applications in a trustworthy and safe manner.
format Online
Article
Text
id pubmed-7274324
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-72743242020-06-05 Evaluation of Uncertainty Quantification in Deep Learning Ståhl, Niclas Falkman, Göran Karlsson, Alexander Mathiason, Gunnar Information Processing and Management of Uncertainty in Knowledge-Based Systems Article Artificial intelligence (AI) is nowadays included into an increasing number of critical systems. Inclusion of AI in such systems may, however, pose a risk, since it is, still, infeasible to build AI systems that know how to function well in situations that differ greatly from what the AI has seen before. Therefore, it is crucial that future AI systems have the ability to not only function well in known domains, but also understand and show when they are uncertain when facing something unknown. In this paper, we evaluate four different methods that have been proposed to correctly quantifying uncertainty when the AI model is faced with new samples. We investigate the behaviour of these models when they are applied to samples far from what these models have seen before, and if they correctly attribute those samples with high uncertainty. We also examine if incorrectly classified samples are attributed with an higher uncertainty than correctly classified samples. The major finding from this simple experiment is, surprisingly, that the evaluated methods capture the uncertainty differently and the correlation between the quantified uncertainty of the models is low. This inconsistency is something that needs to be further understood and solved before AI can be used in critical applications in a trustworthy and safe manner. 2020-05-18 /pmc/articles/PMC7274324/ http://dx.doi.org/10.1007/978-3-030-50146-4_41 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Ståhl, Niclas
Falkman, Göran
Karlsson, Alexander
Mathiason, Gunnar
Evaluation of Uncertainty Quantification in Deep Learning
title Evaluation of Uncertainty Quantification in Deep Learning
title_full Evaluation of Uncertainty Quantification in Deep Learning
title_fullStr Evaluation of Uncertainty Quantification in Deep Learning
title_full_unstemmed Evaluation of Uncertainty Quantification in Deep Learning
title_short Evaluation of Uncertainty Quantification in Deep Learning
title_sort evaluation of uncertainty quantification in deep learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7274324/
http://dx.doi.org/10.1007/978-3-030-50146-4_41
work_keys_str_mv AT stahlniclas evaluationofuncertaintyquantificationindeeplearning
AT falkmangoran evaluationofuncertaintyquantificationindeeplearning
AT karlssonalexander evaluationofuncertaintyquantificationindeeplearning
AT mathiasongunnar evaluationofuncertaintyquantificationindeeplearning