Cargando…

Transformer-based deep learning for predicting protein properties in the life sciences

Recent developments in deep learning, coupled with an increasing number of sequenced proteins, have led to a breakthrough in life science applications, in particular in protein property prediction. There is hope that deep learning can close the gap between the number of sequenced proteins and protei...

Descripción completa

Detalles Bibliográficos
Autores principales: Chandra, Abel, Tünnermann, Laura, Löfstedt, Tommy, Gratz, Regina
Formato: Online Artículo Texto
Lenguaje:English
Publicado: eLife Sciences Publications, Ltd 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9848389/
https://www.ncbi.nlm.nih.gov/pubmed/36651724
http://dx.doi.org/10.7554/eLife.82819
_version_ 1784871698009423872
author Chandra, Abel
Tünnermann, Laura
Löfstedt, Tommy
Gratz, Regina
author_facet Chandra, Abel
Tünnermann, Laura
Löfstedt, Tommy
Gratz, Regina
author_sort Chandra, Abel
collection PubMed
description Recent developments in deep learning, coupled with an increasing number of sequenced proteins, have led to a breakthrough in life science applications, in particular in protein property prediction. There is hope that deep learning can close the gap between the number of sequenced proteins and proteins with known properties based on lab experiments. Language models from the field of natural language processing have gained popularity for protein property predictions and have led to a new computational revolution in biology, where old prediction results are being improved regularly. Such models can learn useful multipurpose representations of proteins from large open repositories of protein sequences and can be used, for instance, to predict protein properties. The field of natural language processing is growing quickly because of developments in a class of models based on a particular model—the Transformer model. We review recent developments and the use of large-scale Transformer models in applications for predicting protein characteristics and how such models can be used to predict, for example, post-translational modifications. We review shortcomings of other deep learning models and explain how the Transformer models have quickly proven to be a very promising way to unravel information hidden in the sequences of amino acids.
format Online
Article
Text
id pubmed-9848389
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher eLife Sciences Publications, Ltd
record_format MEDLINE/PubMed
spelling pubmed-98483892023-01-19 Transformer-based deep learning for predicting protein properties in the life sciences Chandra, Abel Tünnermann, Laura Löfstedt, Tommy Gratz, Regina eLife Computational and Systems Biology Recent developments in deep learning, coupled with an increasing number of sequenced proteins, have led to a breakthrough in life science applications, in particular in protein property prediction. There is hope that deep learning can close the gap between the number of sequenced proteins and proteins with known properties based on lab experiments. Language models from the field of natural language processing have gained popularity for protein property predictions and have led to a new computational revolution in biology, where old prediction results are being improved regularly. Such models can learn useful multipurpose representations of proteins from large open repositories of protein sequences and can be used, for instance, to predict protein properties. The field of natural language processing is growing quickly because of developments in a class of models based on a particular model—the Transformer model. We review recent developments and the use of large-scale Transformer models in applications for predicting protein characteristics and how such models can be used to predict, for example, post-translational modifications. We review shortcomings of other deep learning models and explain how the Transformer models have quickly proven to be a very promising way to unravel information hidden in the sequences of amino acids. eLife Sciences Publications, Ltd 2023-01-18 /pmc/articles/PMC9848389/ /pubmed/36651724 http://dx.doi.org/10.7554/eLife.82819 Text en © 2023, Chandra et al https://creativecommons.org/licenses/by/4.0/This article is distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited.
spellingShingle Computational and Systems Biology
Chandra, Abel
Tünnermann, Laura
Löfstedt, Tommy
Gratz, Regina
Transformer-based deep learning for predicting protein properties in the life sciences
title Transformer-based deep learning for predicting protein properties in the life sciences
title_full Transformer-based deep learning for predicting protein properties in the life sciences
title_fullStr Transformer-based deep learning for predicting protein properties in the life sciences
title_full_unstemmed Transformer-based deep learning for predicting protein properties in the life sciences
title_short Transformer-based deep learning for predicting protein properties in the life sciences
title_sort transformer-based deep learning for predicting protein properties in the life sciences
topic Computational and Systems Biology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9848389/
https://www.ncbi.nlm.nih.gov/pubmed/36651724
http://dx.doi.org/10.7554/eLife.82819
work_keys_str_mv AT chandraabel transformerbaseddeeplearningforpredictingproteinpropertiesinthelifesciences
AT tunnermannlaura transformerbaseddeeplearningforpredictingproteinpropertiesinthelifesciences
AT lofstedttommy transformerbaseddeeplearningforpredictingproteinpropertiesinthelifesciences
AT gratzregina transformerbaseddeeplearningforpredictingproteinpropertiesinthelifesciences