Cargando…
BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning
MOTIVATION: Increasingly comprehensive characterization of cancer-associated genetic alterations has paved the way for the development of highly specific therapeutic vaccines. Predicting precisely the binding and presentation of peptides to major histocompatibility complex (MHC) alleles is an import...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9502151/ https://www.ncbi.nlm.nih.gov/pubmed/34096999 http://dx.doi.org/10.1093/bioinformatics/btab422 |
_version_ | 1784795635919093760 |
---|---|
author | Cheng, Jun Bendjama, Kaïdre Rittner, Karola Malone, Brandon |
author_facet | Cheng, Jun Bendjama, Kaïdre Rittner, Karola Malone, Brandon |
author_sort | Cheng, Jun |
collection | PubMed |
description | MOTIVATION: Increasingly comprehensive characterization of cancer-associated genetic alterations has paved the way for the development of highly specific therapeutic vaccines. Predicting precisely the binding and presentation of peptides to major histocompatibility complex (MHC) alleles is an important step toward such therapies. Recent data suggest that presentation of both class I and II epitopes are critical for the induction of a sustained effective immune response. However, the prediction performance for MHC class II has been limited compared to class I. RESULTS: We present a transformer neural network model which leverages self-supervised pretraining from a large corpus of protein sequences. We also propose a multiple instance learning (MIL) framework to deconvolve mass spectrometry data where multiple potential MHC alleles may have presented each peptide. We show that pretraining boosted the performance for these tasks. Combining pretraining and the novel MIL approach, our model outperforms state-of-the-art models based on peptide and MHC sequence only for both binding and cell surface presentation predictions. AVAILABILITY AND IMPLEMENTATION: Our source code is available at https://github.com/s6juncheng/BERTMHC under a noncommercial license. A webserver is available at https://bertmhc.privacy.nlehd.de/ SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online. |
format | Online Article Text |
id | pubmed-9502151 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-95021512022-09-26 BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning Cheng, Jun Bendjama, Kaïdre Rittner, Karola Malone, Brandon Bioinformatics Original Papers MOTIVATION: Increasingly comprehensive characterization of cancer-associated genetic alterations has paved the way for the development of highly specific therapeutic vaccines. Predicting precisely the binding and presentation of peptides to major histocompatibility complex (MHC) alleles is an important step toward such therapies. Recent data suggest that presentation of both class I and II epitopes are critical for the induction of a sustained effective immune response. However, the prediction performance for MHC class II has been limited compared to class I. RESULTS: We present a transformer neural network model which leverages self-supervised pretraining from a large corpus of protein sequences. We also propose a multiple instance learning (MIL) framework to deconvolve mass spectrometry data where multiple potential MHC alleles may have presented each peptide. We show that pretraining boosted the performance for these tasks. Combining pretraining and the novel MIL approach, our model outperforms state-of-the-art models based on peptide and MHC sequence only for both binding and cell surface presentation predictions. AVAILABILITY AND IMPLEMENTATION: Our source code is available at https://github.com/s6juncheng/BERTMHC under a noncommercial license. A webserver is available at https://bertmhc.privacy.nlehd.de/ SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online. Oxford University Press 2021-06-07 /pmc/articles/PMC9502151/ /pubmed/34096999 http://dx.doi.org/10.1093/bioinformatics/btab422 Text en © The Author(s) 2021. Published by Oxford University Press. https://creativecommons.org/licenses/by-nc/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com |
spellingShingle | Original Papers Cheng, Jun Bendjama, Kaïdre Rittner, Karola Malone, Brandon BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning |
title | BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning |
title_full | BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning |
title_fullStr | BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning |
title_full_unstemmed | BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning |
title_short | BERTMHC: improved MHC–peptide class II interaction prediction with transformer and multiple instance learning |
title_sort | bertmhc: improved mhc–peptide class ii interaction prediction with transformer and multiple instance learning |
topic | Original Papers |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9502151/ https://www.ncbi.nlm.nih.gov/pubmed/34096999 http://dx.doi.org/10.1093/bioinformatics/btab422 |
work_keys_str_mv | AT chengjun bertmhcimprovedmhcpeptideclassiiinteractionpredictionwithtransformerandmultipleinstancelearning AT bendjamakaidre bertmhcimprovedmhcpeptideclassiiinteractionpredictionwithtransformerandmultipleinstancelearning AT rittnerkarola bertmhcimprovedmhcpeptideclassiiinteractionpredictionwithtransformerandmultipleinstancelearning AT malonebrandon bertmhcimprovedmhcpeptideclassiiinteractionpredictionwithtransformerandmultipleinstancelearning |