Cargando…
Entropy and Variability: A Second Opinion by Deep Learning
Background: Analysis of the distribution of amino acid types found at equivalent positions in multiple sequence alignments has found applications in human genetics, protein engineering, drug design, protein structure prediction, and many other fields. These analyses tend to revolve around measures o...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9775329/ https://www.ncbi.nlm.nih.gov/pubmed/36551168 http://dx.doi.org/10.3390/biom12121740 |
_version_ | 1784855617680179200 |
---|---|
author | Rademaker, Daniel T. Xue, Li C. ‘t Hoen, Peter A. C. Vriend, Gert |
author_facet | Rademaker, Daniel T. Xue, Li C. ‘t Hoen, Peter A. C. Vriend, Gert |
author_sort | Rademaker, Daniel T. |
collection | PubMed |
description | Background: Analysis of the distribution of amino acid types found at equivalent positions in multiple sequence alignments has found applications in human genetics, protein engineering, drug design, protein structure prediction, and many other fields. These analyses tend to revolve around measures of the distribution of the twenty amino acid types found at evolutionary equivalent positions: the columns in multiple sequence alignments. Commonly used measures are variability, average hydrophobicity, or Shannon entropy. One of these techniques, called entropy–variability analysis, as the name already suggests, reduces the distribution of observed residue types in one column to two numbers: the Shannon entropy and the variability as defined by the number of residue types observed. Results: We applied a deep learning, unsupervised feature extraction method to analyse the multiple sequence alignments of all human proteins. An auto-encoder neural architecture was trained on 27,835 multiple sequence alignments for human proteins to obtain the two features that best describe the seven million variability patterns. These two unsupervised learned features strongly resemble entropy and variability, indicating that these are the projections that retain most information when reducing the dimensionality of the information hidden in columns in multiple sequence alignments. |
format | Online Article Text |
id | pubmed-9775329 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97753292022-12-23 Entropy and Variability: A Second Opinion by Deep Learning Rademaker, Daniel T. Xue, Li C. ‘t Hoen, Peter A. C. Vriend, Gert Biomolecules Article Background: Analysis of the distribution of amino acid types found at equivalent positions in multiple sequence alignments has found applications in human genetics, protein engineering, drug design, protein structure prediction, and many other fields. These analyses tend to revolve around measures of the distribution of the twenty amino acid types found at evolutionary equivalent positions: the columns in multiple sequence alignments. Commonly used measures are variability, average hydrophobicity, or Shannon entropy. One of these techniques, called entropy–variability analysis, as the name already suggests, reduces the distribution of observed residue types in one column to two numbers: the Shannon entropy and the variability as defined by the number of residue types observed. Results: We applied a deep learning, unsupervised feature extraction method to analyse the multiple sequence alignments of all human proteins. An auto-encoder neural architecture was trained on 27,835 multiple sequence alignments for human proteins to obtain the two features that best describe the seven million variability patterns. These two unsupervised learned features strongly resemble entropy and variability, indicating that these are the projections that retain most information when reducing the dimensionality of the information hidden in columns in multiple sequence alignments. MDPI 2022-11-23 /pmc/articles/PMC9775329/ /pubmed/36551168 http://dx.doi.org/10.3390/biom12121740 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Rademaker, Daniel T. Xue, Li C. ‘t Hoen, Peter A. C. Vriend, Gert Entropy and Variability: A Second Opinion by Deep Learning |
title | Entropy and Variability: A Second Opinion by Deep Learning |
title_full | Entropy and Variability: A Second Opinion by Deep Learning |
title_fullStr | Entropy and Variability: A Second Opinion by Deep Learning |
title_full_unstemmed | Entropy and Variability: A Second Opinion by Deep Learning |
title_short | Entropy and Variability: A Second Opinion by Deep Learning |
title_sort | entropy and variability: a second opinion by deep learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9775329/ https://www.ncbi.nlm.nih.gov/pubmed/36551168 http://dx.doi.org/10.3390/biom12121740 |
work_keys_str_mv | AT rademakerdanielt entropyandvariabilityasecondopinionbydeeplearning AT xuelic entropyandvariabilityasecondopinionbydeeplearning AT thoenpeterac entropyandvariabilityasecondopinionbydeeplearning AT vriendgert entropyandvariabilityasecondopinionbydeeplearning |