Cargando…

Explainable AI: A Review of Machine Learning Interpretability Methods

Recent advances in artificial intelligence (AI) have led to its widespread industrial adoption, with machine learning systems demonstrating superhuman performance in a significant number of tasks. However, this surge in performance, has often been achieved through increased model complexity, turning...

Descripción completa

Detalles Bibliográficos
Autores principales: Linardatos, Pantelis, Papastefanopoulos, Vasilis, Kotsiantis, Sotiris
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7824368/
https://www.ncbi.nlm.nih.gov/pubmed/33375658
http://dx.doi.org/10.3390/e23010018
_version_ 1783640060171649024
author Linardatos, Pantelis
Papastefanopoulos, Vasilis
Kotsiantis, Sotiris
author_facet Linardatos, Pantelis
Papastefanopoulos, Vasilis
Kotsiantis, Sotiris
author_sort Linardatos, Pantelis
collection PubMed
description Recent advances in artificial intelligence (AI) have led to its widespread industrial adoption, with machine learning systems demonstrating superhuman performance in a significant number of tasks. However, this surge in performance, has often been achieved through increased model complexity, turning such systems into “black box” approaches and causing uncertainty regarding the way they operate and, ultimately, the way that they come to decisions. This ambiguity has made it problematic for machine learning systems to be adopted in sensitive yet critical domains, where their value could be immense, such as healthcare. As a result, scientific interest in the field of Explainable Artificial Intelligence (XAI), a field that is concerned with the development of new methods that explain and interpret machine learning models, has been tremendously reignited over recent years. This study focuses on machine learning interpretability methods; more specifically, a literature review and taxonomy of these methods are presented, as well as links to their programming implementations, in the hope that this survey would serve as a reference point for both theorists and practitioners.
format Online
Article
Text
id pubmed-7824368
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-78243682021-02-24 Explainable AI: A Review of Machine Learning Interpretability Methods Linardatos, Pantelis Papastefanopoulos, Vasilis Kotsiantis, Sotiris Entropy (Basel) Review Recent advances in artificial intelligence (AI) have led to its widespread industrial adoption, with machine learning systems demonstrating superhuman performance in a significant number of tasks. However, this surge in performance, has often been achieved through increased model complexity, turning such systems into “black box” approaches and causing uncertainty regarding the way they operate and, ultimately, the way that they come to decisions. This ambiguity has made it problematic for machine learning systems to be adopted in sensitive yet critical domains, where their value could be immense, such as healthcare. As a result, scientific interest in the field of Explainable Artificial Intelligence (XAI), a field that is concerned with the development of new methods that explain and interpret machine learning models, has been tremendously reignited over recent years. This study focuses on machine learning interpretability methods; more specifically, a literature review and taxonomy of these methods are presented, as well as links to their programming implementations, in the hope that this survey would serve as a reference point for both theorists and practitioners. MDPI 2020-12-25 /pmc/articles/PMC7824368/ /pubmed/33375658 http://dx.doi.org/10.3390/e23010018 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Review
Linardatos, Pantelis
Papastefanopoulos, Vasilis
Kotsiantis, Sotiris
Explainable AI: A Review of Machine Learning Interpretability Methods
title Explainable AI: A Review of Machine Learning Interpretability Methods
title_full Explainable AI: A Review of Machine Learning Interpretability Methods
title_fullStr Explainable AI: A Review of Machine Learning Interpretability Methods
title_full_unstemmed Explainable AI: A Review of Machine Learning Interpretability Methods
title_short Explainable AI: A Review of Machine Learning Interpretability Methods
title_sort explainable ai: a review of machine learning interpretability methods
topic Review
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7824368/
https://www.ncbi.nlm.nih.gov/pubmed/33375658
http://dx.doi.org/10.3390/e23010018
work_keys_str_mv AT linardatospantelis explainableaiareviewofmachinelearninginterpretabilitymethods
AT papastefanopoulosvasilis explainableaiareviewofmachinelearninginterpretabilitymethods
AT kotsiantissotiris explainableaiareviewofmachinelearninginterpretabilitymethods