Cargando…

Understanding, Explanation, and Active Inference

While machine learning techniques have been transformative in solving a range of problems, an important challenge is to understand why they arrive at the decisions they output. Some have argued that this necessitates augmenting machine intelligence with understanding such that, when queried, a machi...

Descripción completa

Detalles Bibliográficos
Autores principales: Parr, Thomas, Pezzulo, Giovanni
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8602880/
https://www.ncbi.nlm.nih.gov/pubmed/34803619
http://dx.doi.org/10.3389/fnsys.2021.772641
_version_ 1784601655991336960
author Parr, Thomas
Pezzulo, Giovanni
author_facet Parr, Thomas
Pezzulo, Giovanni
author_sort Parr, Thomas
collection PubMed
description While machine learning techniques have been transformative in solving a range of problems, an important challenge is to understand why they arrive at the decisions they output. Some have argued that this necessitates augmenting machine intelligence with understanding such that, when queried, a machine is able to explain its behaviour (i.e., explainable AI). In this article, we address the issue of machine understanding from the perspective of active inference. This paradigm enables decision making based upon a model of how data are generated. The generative model contains those variables required to explain sensory data, and its inversion may be seen as an attempt to explain the causes of these data. Here we are interested in explanations of one’s own actions. This implies a deep generative model that includes a model of the world, used to infer policies, and a higher-level model that attempts to predict which policies will be selected based upon a space of hypothetical (i.e., counterfactual) explanations—and which can subsequently be used to provide (retrospective) explanations about the policies pursued. We illustrate the construct validity of this notion of understanding in relation to human understanding by highlighting the similarities in computational architecture and the consequences of its dysfunction.
format Online
Article
Text
id pubmed-8602880
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-86028802021-11-20 Understanding, Explanation, and Active Inference Parr, Thomas Pezzulo, Giovanni Front Syst Neurosci Neuroscience While machine learning techniques have been transformative in solving a range of problems, an important challenge is to understand why they arrive at the decisions they output. Some have argued that this necessitates augmenting machine intelligence with understanding such that, when queried, a machine is able to explain its behaviour (i.e., explainable AI). In this article, we address the issue of machine understanding from the perspective of active inference. This paradigm enables decision making based upon a model of how data are generated. The generative model contains those variables required to explain sensory data, and its inversion may be seen as an attempt to explain the causes of these data. Here we are interested in explanations of one’s own actions. This implies a deep generative model that includes a model of the world, used to infer policies, and a higher-level model that attempts to predict which policies will be selected based upon a space of hypothetical (i.e., counterfactual) explanations—and which can subsequently be used to provide (retrospective) explanations about the policies pursued. We illustrate the construct validity of this notion of understanding in relation to human understanding by highlighting the similarities in computational architecture and the consequences of its dysfunction. Frontiers Media S.A. 2021-11-05 /pmc/articles/PMC8602880/ /pubmed/34803619 http://dx.doi.org/10.3389/fnsys.2021.772641 Text en Copyright © 2021 Parr and Pezzulo. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Parr, Thomas
Pezzulo, Giovanni
Understanding, Explanation, and Active Inference
title Understanding, Explanation, and Active Inference
title_full Understanding, Explanation, and Active Inference
title_fullStr Understanding, Explanation, and Active Inference
title_full_unstemmed Understanding, Explanation, and Active Inference
title_short Understanding, Explanation, and Active Inference
title_sort understanding, explanation, and active inference
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8602880/
https://www.ncbi.nlm.nih.gov/pubmed/34803619
http://dx.doi.org/10.3389/fnsys.2021.772641
work_keys_str_mv AT parrthomas understandingexplanationandactiveinference
AT pezzulogiovanni understandingexplanationandactiveinference