Cargando…

Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach

Model understanding is critical in many domains, particularly those involved in high-stakes decisions, e.g., medicine, criminal justice, and autonomous driving. Explainable AI (XAI) methods are essential for working with black-box models such as convolutional neural networks. This paper evaluates th...

Descripción completa

Detalles Bibliográficos
Autores principales: Sudars, Kaspars, Namatēvs, Ivars, Ozols, Kaspars
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8874740/
https://www.ncbi.nlm.nih.gov/pubmed/35200732
http://dx.doi.org/10.3390/jimaging8020030
_version_ 1784657761301168128
author Sudars, Kaspars
Namatēvs, Ivars
Ozols, Kaspars
author_facet Sudars, Kaspars
Namatēvs, Ivars
Ozols, Kaspars
author_sort Sudars, Kaspars
collection PubMed
description Model understanding is critical in many domains, particularly those involved in high-stakes decisions, e.g., medicine, criminal justice, and autonomous driving. Explainable AI (XAI) methods are essential for working with black-box models such as convolutional neural networks. This paper evaluates the traffic sign classifier of the Deep Neural Network (DNN) from the Programmable Systems for Intelligence in Automobiles (PRYSTINE) project for explainability. The results of explanations were further used for the CNN PRYSTINE classifier vague kernels’ compression. Then, the precision of the classifier was evaluated in different pruning scenarios. The proposed classifier performance methodology was realised by creating an original traffic sign and traffic light classification and explanation code. First, the status of the kernels of the network was evaluated for explainability. For this task, the post-hoc, local, meaningful perturbation-based forward explainable method was integrated into the model to evaluate each kernel status of the network. This method enabled distinguishing high- and low-impact kernels in the CNN. Second, the vague kernels of the classifier of the last layer before the fully connected layer were excluded by withdrawing them from the network. Third, the network’s precision was evaluated in different kernel compression levels. It is shown that by using the XAI approach for network kernel compression, the pruning of 5% of kernels leads to a 2% loss in traffic sign and traffic light classification precision. The proposed methodology is crucial where execution time and processing capacity prevail.
format Online
Article
Text
id pubmed-8874740
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-88747402022-02-26 Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach Sudars, Kaspars Namatēvs, Ivars Ozols, Kaspars J Imaging Article Model understanding is critical in many domains, particularly those involved in high-stakes decisions, e.g., medicine, criminal justice, and autonomous driving. Explainable AI (XAI) methods are essential for working with black-box models such as convolutional neural networks. This paper evaluates the traffic sign classifier of the Deep Neural Network (DNN) from the Programmable Systems for Intelligence in Automobiles (PRYSTINE) project for explainability. The results of explanations were further used for the CNN PRYSTINE classifier vague kernels’ compression. Then, the precision of the classifier was evaluated in different pruning scenarios. The proposed classifier performance methodology was realised by creating an original traffic sign and traffic light classification and explanation code. First, the status of the kernels of the network was evaluated for explainability. For this task, the post-hoc, local, meaningful perturbation-based forward explainable method was integrated into the model to evaluate each kernel status of the network. This method enabled distinguishing high- and low-impact kernels in the CNN. Second, the vague kernels of the classifier of the last layer before the fully connected layer were excluded by withdrawing them from the network. Third, the network’s precision was evaluated in different kernel compression levels. It is shown that by using the XAI approach for network kernel compression, the pruning of 5% of kernels leads to a 2% loss in traffic sign and traffic light classification precision. The proposed methodology is crucial where execution time and processing capacity prevail. MDPI 2022-01-30 /pmc/articles/PMC8874740/ /pubmed/35200732 http://dx.doi.org/10.3390/jimaging8020030 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sudars, Kaspars
Namatēvs, Ivars
Ozols, Kaspars
Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach
title Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach
title_full Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach
title_fullStr Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach
title_full_unstemmed Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach
title_short Improving Performance of the PRYSTINE Traffic Sign Classification by Using a Perturbation-Based Explainability Approach
title_sort improving performance of the prystine traffic sign classification by using a perturbation-based explainability approach
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8874740/
https://www.ncbi.nlm.nih.gov/pubmed/35200732
http://dx.doi.org/10.3390/jimaging8020030
work_keys_str_mv AT sudarskaspars improvingperformanceoftheprystinetrafficsignclassificationbyusingaperturbationbasedexplainabilityapproach
AT namatevsivars improvingperformanceoftheprystinetrafficsignclassificationbyusingaperturbationbasedexplainabilityapproach
AT ozolskaspars improvingperformanceoftheprystinetrafficsignclassificationbyusingaperturbationbasedexplainabilityapproach