Cargando…
Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks
Deep neural networks have demonstrated improved performance at predicting the sequence specificities of DNA- and RNA-binding proteins compared to previous methods that rely on k-mers and position weight matrices. To gain insights into why a DNN makes a given prediction, model interpretability method...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8118286/ https://www.ncbi.nlm.nih.gov/pubmed/33983921 http://dx.doi.org/10.1371/journal.pcbi.1008925 |
_version_ | 1783691717762875392 |
---|---|
author | Koo, Peter K. Majdandzic, Antonio Ploenzke, Matthew Anand, Praveen Paul, Steffan B. |
author_facet | Koo, Peter K. Majdandzic, Antonio Ploenzke, Matthew Anand, Praveen Paul, Steffan B. |
author_sort | Koo, Peter K. |
collection | PubMed |
description | Deep neural networks have demonstrated improved performance at predicting the sequence specificities of DNA- and RNA-binding proteins compared to previous methods that rely on k-mers and position weight matrices. To gain insights into why a DNN makes a given prediction, model interpretability methods, such as attribution methods, can be employed to identify motif-like representations along a given sequence. Because explanations are given on an individual sequence basis and can vary substantially across sequences, deducing generalizable trends across the dataset and quantifying their effect size remains a challenge. Here we introduce global importance analysis (GIA), a model interpretability method that quantifies the population-level effect size that putative patterns have on model predictions. GIA provides an avenue to quantitatively test hypotheses of putative patterns and their interactions with other patterns, as well as map out specific functions the network has learned. As a case study, we demonstrate the utility of GIA on the computational task of predicting RNA-protein interactions from sequence. We first introduce a convolutional network, we call ResidualBind, and benchmark its performance against previous methods on RNAcompete data. Using GIA, we then demonstrate that in addition to sequence motifs, ResidualBind learns a model that considers the number of motifs, their spacing, and sequence context, such as RNA secondary structure and GC-bias. |
format | Online Article Text |
id | pubmed-8118286 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-81182862021-05-24 Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks Koo, Peter K. Majdandzic, Antonio Ploenzke, Matthew Anand, Praveen Paul, Steffan B. PLoS Comput Biol Research Article Deep neural networks have demonstrated improved performance at predicting the sequence specificities of DNA- and RNA-binding proteins compared to previous methods that rely on k-mers and position weight matrices. To gain insights into why a DNN makes a given prediction, model interpretability methods, such as attribution methods, can be employed to identify motif-like representations along a given sequence. Because explanations are given on an individual sequence basis and can vary substantially across sequences, deducing generalizable trends across the dataset and quantifying their effect size remains a challenge. Here we introduce global importance analysis (GIA), a model interpretability method that quantifies the population-level effect size that putative patterns have on model predictions. GIA provides an avenue to quantitatively test hypotheses of putative patterns and their interactions with other patterns, as well as map out specific functions the network has learned. As a case study, we demonstrate the utility of GIA on the computational task of predicting RNA-protein interactions from sequence. We first introduce a convolutional network, we call ResidualBind, and benchmark its performance against previous methods on RNAcompete data. Using GIA, we then demonstrate that in addition to sequence motifs, ResidualBind learns a model that considers the number of motifs, their spacing, and sequence context, such as RNA secondary structure and GC-bias. Public Library of Science 2021-05-13 /pmc/articles/PMC8118286/ /pubmed/33983921 http://dx.doi.org/10.1371/journal.pcbi.1008925 Text en © 2021 Koo et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Koo, Peter K. Majdandzic, Antonio Ploenzke, Matthew Anand, Praveen Paul, Steffan B. Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks |
title | Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks |
title_full | Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks |
title_fullStr | Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks |
title_full_unstemmed | Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks |
title_short | Global importance analysis: An interpretability method to quantify importance of genomic features in deep neural networks |
title_sort | global importance analysis: an interpretability method to quantify importance of genomic features in deep neural networks |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8118286/ https://www.ncbi.nlm.nih.gov/pubmed/33983921 http://dx.doi.org/10.1371/journal.pcbi.1008925 |
work_keys_str_mv | AT koopeterk globalimportanceanalysisaninterpretabilitymethodtoquantifyimportanceofgenomicfeaturesindeepneuralnetworks AT majdandzicantonio globalimportanceanalysisaninterpretabilitymethodtoquantifyimportanceofgenomicfeaturesindeepneuralnetworks AT ploenzkematthew globalimportanceanalysisaninterpretabilitymethodtoquantifyimportanceofgenomicfeaturesindeepneuralnetworks AT anandpraveen globalimportanceanalysisaninterpretabilitymethodtoquantifyimportanceofgenomicfeaturesindeepneuralnetworks AT paulsteffanb globalimportanceanalysisaninterpretabilitymethodtoquantifyimportanceofgenomicfeaturesindeepneuralnetworks |