Cargando…
Mitigating belief projection in explainable artificial intelligence via Bayesian teaching
State-of-the-art deep-learning systems use decision rules that are challenging for humans to model. Explainable AI (XAI) attempts to improve human understanding but rarely accounts for how people typically reason about unfamiliar agents. We propose explicitly modelling the human explainee via Bayesi...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8110978/ https://www.ncbi.nlm.nih.gov/pubmed/33972625 http://dx.doi.org/10.1038/s41598-021-89267-4 |
_version_ | 1783690404811505664 |
---|---|
author | Yang, Scott Cheng-Hsin Vong, Wai Keen Sojitra, Ravi B. Folke, Tomas Shafto, Patrick |
author_facet | Yang, Scott Cheng-Hsin Vong, Wai Keen Sojitra, Ravi B. Folke, Tomas Shafto, Patrick |
author_sort | Yang, Scott Cheng-Hsin |
collection | PubMed |
description | State-of-the-art deep-learning systems use decision rules that are challenging for humans to model. Explainable AI (XAI) attempts to improve human understanding but rarely accounts for how people typically reason about unfamiliar agents. We propose explicitly modelling the human explainee via Bayesian teaching, which evaluates explanations by how much they shift explainees’ inferences toward a desired goal. We assess Bayesian teaching in a binary image classification task across a variety of contexts. Absent intervention, participants predict that the AI’s classifications will match their own, but explanations generated by Bayesian teaching improve their ability to predict the AI’s judgements by moving them away from this prior belief. Bayesian teaching further allows each case to be broken down into sub-examples (here saliency maps). These sub-examples complement whole examples by improving error detection for familiar categories, whereas whole examples help predict correct AI judgements of unfamiliar cases. |
format | Online Article Text |
id | pubmed-8110978 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-81109782021-05-12 Mitigating belief projection in explainable artificial intelligence via Bayesian teaching Yang, Scott Cheng-Hsin Vong, Wai Keen Sojitra, Ravi B. Folke, Tomas Shafto, Patrick Sci Rep Article State-of-the-art deep-learning systems use decision rules that are challenging for humans to model. Explainable AI (XAI) attempts to improve human understanding but rarely accounts for how people typically reason about unfamiliar agents. We propose explicitly modelling the human explainee via Bayesian teaching, which evaluates explanations by how much they shift explainees’ inferences toward a desired goal. We assess Bayesian teaching in a binary image classification task across a variety of contexts. Absent intervention, participants predict that the AI’s classifications will match their own, but explanations generated by Bayesian teaching improve their ability to predict the AI’s judgements by moving them away from this prior belief. Bayesian teaching further allows each case to be broken down into sub-examples (here saliency maps). These sub-examples complement whole examples by improving error detection for familiar categories, whereas whole examples help predict correct AI judgements of unfamiliar cases. Nature Publishing Group UK 2021-05-10 /pmc/articles/PMC8110978/ /pubmed/33972625 http://dx.doi.org/10.1038/s41598-021-89267-4 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author (s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Yang, Scott Cheng-Hsin Vong, Wai Keen Sojitra, Ravi B. Folke, Tomas Shafto, Patrick Mitigating belief projection in explainable artificial intelligence via Bayesian teaching |
title | Mitigating belief projection in explainable artificial intelligence via Bayesian teaching |
title_full | Mitigating belief projection in explainable artificial intelligence via Bayesian teaching |
title_fullStr | Mitigating belief projection in explainable artificial intelligence via Bayesian teaching |
title_full_unstemmed | Mitigating belief projection in explainable artificial intelligence via Bayesian teaching |
title_short | Mitigating belief projection in explainable artificial intelligence via Bayesian teaching |
title_sort | mitigating belief projection in explainable artificial intelligence via bayesian teaching |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8110978/ https://www.ncbi.nlm.nih.gov/pubmed/33972625 http://dx.doi.org/10.1038/s41598-021-89267-4 |
work_keys_str_mv | AT yangscottchenghsin mitigatingbeliefprojectioninexplainableartificialintelligenceviabayesianteaching AT vongwaikeen mitigatingbeliefprojectioninexplainableartificialintelligenceviabayesianteaching AT sojitraravib mitigatingbeliefprojectioninexplainableartificialintelligenceviabayesianteaching AT folketomas mitigatingbeliefprojectioninexplainableartificialintelligenceviabayesianteaching AT shaftopatrick mitigatingbeliefprojectioninexplainableartificialintelligenceviabayesianteaching |