Cargando…
Evaluation of Uncertainty Quantification in Deep Learning
Artificial intelligence (AI) is nowadays included into an increasing number of critical systems. Inclusion of AI in such systems may, however, pose a risk, since it is, still, infeasible to build AI systems that know how to function well in situations that differ greatly from what the AI has seen be...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7274324/ http://dx.doi.org/10.1007/978-3-030-50146-4_41 |
_version_ | 1783542556547612672 |
---|---|
author | Ståhl, Niclas Falkman, Göran Karlsson, Alexander Mathiason, Gunnar |
author_facet | Ståhl, Niclas Falkman, Göran Karlsson, Alexander Mathiason, Gunnar |
author_sort | Ståhl, Niclas |
collection | PubMed |
description | Artificial intelligence (AI) is nowadays included into an increasing number of critical systems. Inclusion of AI in such systems may, however, pose a risk, since it is, still, infeasible to build AI systems that know how to function well in situations that differ greatly from what the AI has seen before. Therefore, it is crucial that future AI systems have the ability to not only function well in known domains, but also understand and show when they are uncertain when facing something unknown. In this paper, we evaluate four different methods that have been proposed to correctly quantifying uncertainty when the AI model is faced with new samples. We investigate the behaviour of these models when they are applied to samples far from what these models have seen before, and if they correctly attribute those samples with high uncertainty. We also examine if incorrectly classified samples are attributed with an higher uncertainty than correctly classified samples. The major finding from this simple experiment is, surprisingly, that the evaluated methods capture the uncertainty differently and the correlation between the quantified uncertainty of the models is low. This inconsistency is something that needs to be further understood and solved before AI can be used in critical applications in a trustworthy and safe manner. |
format | Online Article Text |
id | pubmed-7274324 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72743242020-06-05 Evaluation of Uncertainty Quantification in Deep Learning Ståhl, Niclas Falkman, Göran Karlsson, Alexander Mathiason, Gunnar Information Processing and Management of Uncertainty in Knowledge-Based Systems Article Artificial intelligence (AI) is nowadays included into an increasing number of critical systems. Inclusion of AI in such systems may, however, pose a risk, since it is, still, infeasible to build AI systems that know how to function well in situations that differ greatly from what the AI has seen before. Therefore, it is crucial that future AI systems have the ability to not only function well in known domains, but also understand and show when they are uncertain when facing something unknown. In this paper, we evaluate four different methods that have been proposed to correctly quantifying uncertainty when the AI model is faced with new samples. We investigate the behaviour of these models when they are applied to samples far from what these models have seen before, and if they correctly attribute those samples with high uncertainty. We also examine if incorrectly classified samples are attributed with an higher uncertainty than correctly classified samples. The major finding from this simple experiment is, surprisingly, that the evaluated methods capture the uncertainty differently and the correlation between the quantified uncertainty of the models is low. This inconsistency is something that needs to be further understood and solved before AI can be used in critical applications in a trustworthy and safe manner. 2020-05-18 /pmc/articles/PMC7274324/ http://dx.doi.org/10.1007/978-3-030-50146-4_41 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Ståhl, Niclas Falkman, Göran Karlsson, Alexander Mathiason, Gunnar Evaluation of Uncertainty Quantification in Deep Learning |
title | Evaluation of Uncertainty Quantification in Deep Learning |
title_full | Evaluation of Uncertainty Quantification in Deep Learning |
title_fullStr | Evaluation of Uncertainty Quantification in Deep Learning |
title_full_unstemmed | Evaluation of Uncertainty Quantification in Deep Learning |
title_short | Evaluation of Uncertainty Quantification in Deep Learning |
title_sort | evaluation of uncertainty quantification in deep learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7274324/ http://dx.doi.org/10.1007/978-3-030-50146-4_41 |
work_keys_str_mv | AT stahlniclas evaluationofuncertaintyquantificationindeeplearning AT falkmangoran evaluationofuncertaintyquantificationindeeplearning AT karlssonalexander evaluationofuncertaintyquantificationindeeplearning AT mathiasongunnar evaluationofuncertaintyquantificationindeeplearning |