Cargando…

Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )

AIMS: Deep neural networks (DNNs) perform excellently in interpreting electrocardiograms (ECGs), both for conventional ECG interpretation and for novel applications such as detection of reduced ejection fraction (EF). Despite these promising developments, implementation is hampered by the lack of tr...

Descripción completa

Detalles Bibliográficos
Autores principales: van de Leur, Rutger R, Bos, Max N, Taha, Karim, Sammani, Arjan, Yeung, Ming Wai, van Duijvenboden, Stefan, Lambiase, Pier D, Hassink, Rutger J, van der Harst, Pim, Doevendans, Pieter A, Gupta, Deepak K, van Es, René
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9707974/
https://www.ncbi.nlm.nih.gov/pubmed/36712164
http://dx.doi.org/10.1093/ehjdh/ztac038
_version_ 1784840819341000704
author van de Leur, Rutger R
Bos, Max N
Taha, Karim
Sammani, Arjan
Yeung, Ming Wai
van Duijvenboden, Stefan
Lambiase, Pier D
Hassink, Rutger J
van der Harst, Pim
Doevendans, Pieter A
Gupta, Deepak K
van Es, René
author_facet van de Leur, Rutger R
Bos, Max N
Taha, Karim
Sammani, Arjan
Yeung, Ming Wai
van Duijvenboden, Stefan
Lambiase, Pier D
Hassink, Rutger J
van der Harst, Pim
Doevendans, Pieter A
Gupta, Deepak K
van Es, René
author_sort van de Leur, Rutger R
collection PubMed
description AIMS: Deep neural networks (DNNs) perform excellently in interpreting electrocardiograms (ECGs), both for conventional ECG interpretation and for novel applications such as detection of reduced ejection fraction (EF). Despite these promising developments, implementation is hampered by the lack of trustworthy techniques to explain the algorithms to clinicians. Especially, currently employed heatmap-based methods have shown to be inaccurate. METHODS AND RESULTS: We present a novel pipeline consisting of a variational auto-encoder (VAE) to learn the underlying factors of variation of the median beat ECG morphology (the FactorECG), which are subsequently used in common and interpretable prediction models. As the ECG factors can be made explainable by generating and visualizing ECGs on both the model and individual level, the pipeline provides improved explainability over heatmap-based methods. By training on a database with 1.1 million ECGs, the VAE can compress the ECG into 21 generative ECG factors, most of which are associated with physiologically valid underlying processes. Performance of the explainable pipeline was similar to ‘black box’ DNNs in conventional ECG interpretation [area under the receiver operating curve (AUROC) 0.94 vs. 0.96], detection of reduced EF (AUROC 0.90 vs. 0.91), and prediction of 1-year mortality (AUROC 0.76 vs. 0.75). Contrary to the ‘black box’ DNNs, our pipeline provided explainability on which morphological ECG changes were important for prediction. Results were confirmed in a population-based external validation dataset. CONCLUSIONS: Future studies on DNNs for ECGs should employ pipelines that are explainable to facilitate clinical implementation by gaining confidence in artificial intelligence and making it possible to identify biased models.
format Online
Article
Text
id pubmed-9707974
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-97079742023-01-27 Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( ) van de Leur, Rutger R Bos, Max N Taha, Karim Sammani, Arjan Yeung, Ming Wai van Duijvenboden, Stefan Lambiase, Pier D Hassink, Rutger J van der Harst, Pim Doevendans, Pieter A Gupta, Deepak K van Es, René Eur Heart J Digit Health Original Article AIMS: Deep neural networks (DNNs) perform excellently in interpreting electrocardiograms (ECGs), both for conventional ECG interpretation and for novel applications such as detection of reduced ejection fraction (EF). Despite these promising developments, implementation is hampered by the lack of trustworthy techniques to explain the algorithms to clinicians. Especially, currently employed heatmap-based methods have shown to be inaccurate. METHODS AND RESULTS: We present a novel pipeline consisting of a variational auto-encoder (VAE) to learn the underlying factors of variation of the median beat ECG morphology (the FactorECG), which are subsequently used in common and interpretable prediction models. As the ECG factors can be made explainable by generating and visualizing ECGs on both the model and individual level, the pipeline provides improved explainability over heatmap-based methods. By training on a database with 1.1 million ECGs, the VAE can compress the ECG into 21 generative ECG factors, most of which are associated with physiologically valid underlying processes. Performance of the explainable pipeline was similar to ‘black box’ DNNs in conventional ECG interpretation [area under the receiver operating curve (AUROC) 0.94 vs. 0.96], detection of reduced EF (AUROC 0.90 vs. 0.91), and prediction of 1-year mortality (AUROC 0.76 vs. 0.75). Contrary to the ‘black box’ DNNs, our pipeline provided explainability on which morphological ECG changes were important for prediction. Results were confirmed in a population-based external validation dataset. CONCLUSIONS: Future studies on DNNs for ECGs should employ pipelines that are explainable to facilitate clinical implementation by gaining confidence in artificial intelligence and making it possible to identify biased models. Oxford University Press 2022-07-25 /pmc/articles/PMC9707974/ /pubmed/36712164 http://dx.doi.org/10.1093/ehjdh/ztac038 Text en © The Author(s) 2022. Published by Oxford University Press on behalf of the European Society of Cardiology. https://creativecommons.org/licenses/by-nc/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com
spellingShingle Original Article
van de Leur, Rutger R
Bos, Max N
Taha, Karim
Sammani, Arjan
Yeung, Ming Wai
van Duijvenboden, Stefan
Lambiase, Pier D
Hassink, Rutger J
van der Harst, Pim
Doevendans, Pieter A
Gupta, Deepak K
van Es, René
Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
title Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
title_full Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
title_fullStr Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
title_full_unstemmed Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
title_short Improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
title_sort improving explainability of deep neural network-based electrocardiogram interpretation using variational auto-encoders( )
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9707974/
https://www.ncbi.nlm.nih.gov/pubmed/36712164
http://dx.doi.org/10.1093/ehjdh/ztac038
work_keys_str_mv AT vandeleurrutgerr improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT bosmaxn improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT tahakarim improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT sammaniarjan improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT yeungmingwai improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT vanduijvenbodenstefan improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT lambiasepierd improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT hassinkrutgerj improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT vanderharstpim improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT doevendanspietera improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT guptadeepakk improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders
AT vanesrene improvingexplainabilityofdeepneuralnetworkbasedelectrocardiograminterpretationusingvariationalautoencoders