Cargando…
SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk
In credit risk estimation, the most important element is obtaining a probability of default as close as possible to the effective risk. This effort quickly prompted new, powerful algorithms that reach a far higher accuracy, but at the cost of losing intelligibility, such as Gradient Boosting or ense...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8484963/ https://www.ncbi.nlm.nih.gov/pubmed/34604738 http://dx.doi.org/10.3389/frai.2021.752558 |
_version_ | 1784577437248520192 |
---|---|
author | Gramegna, Alex Giudici, Paolo |
author_facet | Gramegna, Alex Giudici, Paolo |
author_sort | Gramegna, Alex |
collection | PubMed |
description | In credit risk estimation, the most important element is obtaining a probability of default as close as possible to the effective risk. This effort quickly prompted new, powerful algorithms that reach a far higher accuracy, but at the cost of losing intelligibility, such as Gradient Boosting or ensemble methods. These models are usually referred to as “black-boxes”, implying that you know the inputs and the output, but there is little way to understand what is going on under the hood. As a response to that, we have seen several different Explainable AI models flourish in recent years, with the aim of letting the user see why the black-box gave a certain output. In this context, we evaluate two very popular eXplainable AI (XAI) models in their ability to discriminate observations into groups, through the application of both unsupervised and predictive modeling to the weights these XAI models assign to features locally. The evaluation is carried out on real Small and Medium Enterprises data, obtained from official italian repositories, and may form the basis for the employment of such XAI models for post-processing features extraction. |
format | Online Article Text |
id | pubmed-8484963 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-84849632021-10-02 SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk Gramegna, Alex Giudici, Paolo Front Artif Intell Artificial Intelligence In credit risk estimation, the most important element is obtaining a probability of default as close as possible to the effective risk. This effort quickly prompted new, powerful algorithms that reach a far higher accuracy, but at the cost of losing intelligibility, such as Gradient Boosting or ensemble methods. These models are usually referred to as “black-boxes”, implying that you know the inputs and the output, but there is little way to understand what is going on under the hood. As a response to that, we have seen several different Explainable AI models flourish in recent years, with the aim of letting the user see why the black-box gave a certain output. In this context, we evaluate two very popular eXplainable AI (XAI) models in their ability to discriminate observations into groups, through the application of both unsupervised and predictive modeling to the weights these XAI models assign to features locally. The evaluation is carried out on real Small and Medium Enterprises data, obtained from official italian repositories, and may form the basis for the employment of such XAI models for post-processing features extraction. Frontiers Media S.A. 2021-09-17 /pmc/articles/PMC8484963/ /pubmed/34604738 http://dx.doi.org/10.3389/frai.2021.752558 Text en Copyright © 2021 Gramegna and Giudici. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Gramegna, Alex Giudici, Paolo SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk |
title | SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk |
title_full | SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk |
title_fullStr | SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk |
title_full_unstemmed | SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk |
title_short | SHAP and LIME: An Evaluation of Discriminative Power in Credit Risk |
title_sort | shap and lime: an evaluation of discriminative power in credit risk |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8484963/ https://www.ncbi.nlm.nih.gov/pubmed/34604738 http://dx.doi.org/10.3389/frai.2021.752558 |
work_keys_str_mv | AT gramegnaalex shapandlimeanevaluationofdiscriminativepowerincreditrisk AT giudicipaolo shapandlimeanevaluationofdiscriminativepowerincreditrisk |