Cargando…
Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network
Recurrent neural networks with memory and attention mechanisms are widely used in natural language processing because they can capture short and long term sequential information for diverse tasks. We propose an integrated deep learning model for microbial DNA sequence data, which exploits convolutio...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8496832/ https://www.ncbi.nlm.nih.gov/pubmed/34550967 http://dx.doi.org/10.1371/journal.pcbi.1009345 |
_version_ | 1784579834011189248 |
---|---|
author | Zhao, Zhengqiao Woloszynek, Stephen Agbavor, Felix Mell, Joshua Chang Sokhansanj, Bahrad A. Rosen, Gail L. |
author_facet | Zhao, Zhengqiao Woloszynek, Stephen Agbavor, Felix Mell, Joshua Chang Sokhansanj, Bahrad A. Rosen, Gail L. |
author_sort | Zhao, Zhengqiao |
collection | PubMed |
description | Recurrent neural networks with memory and attention mechanisms are widely used in natural language processing because they can capture short and long term sequential information for diverse tasks. We propose an integrated deep learning model for microbial DNA sequence data, which exploits convolutional neural networks, recurrent neural networks, and attention mechanisms to predict taxonomic classifications and sample-associated attributes, such as the relationship between the microbiome and host phenotype, on the read/sequence level. In this paper, we develop this novel deep learning approach and evaluate its application to amplicon sequences. We apply our approach to short DNA reads and full sequences of 16S ribosomal RNA (rRNA) marker genes, which identify the heterogeneity of a microbial community sample. We demonstrate that our implementation of a novel attention-based deep network architecture, Read2Pheno, achieves read-level phenotypic prediction. Training Read2Pheno models will encode sequences (reads) into dense, meaningful representations: learned embedded vectors output from the intermediate layer of the network model, which can provide biological insight when visualized. The attention layer of Read2Pheno models can also automatically identify nucleotide regions in reads/sequences which are particularly informative for classification. As such, this novel approach can avoid pre/post-processing and manual interpretation required with conventional approaches to microbiome sequence classification. We further show, as proof-of-concept, that aggregating read-level information can robustly predict microbial community properties, host phenotype, and taxonomic classification, with performance at least comparable to conventional approaches. An implementation of the attention-based deep learning network is available at https://github.com/EESI/sequence_attention (a python package) and https://github.com/EESI/seq2att (a command line tool). |
format | Online Article Text |
id | pubmed-8496832 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-84968322021-10-08 Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network Zhao, Zhengqiao Woloszynek, Stephen Agbavor, Felix Mell, Joshua Chang Sokhansanj, Bahrad A. Rosen, Gail L. PLoS Comput Biol Research Article Recurrent neural networks with memory and attention mechanisms are widely used in natural language processing because they can capture short and long term sequential information for diverse tasks. We propose an integrated deep learning model for microbial DNA sequence data, which exploits convolutional neural networks, recurrent neural networks, and attention mechanisms to predict taxonomic classifications and sample-associated attributes, such as the relationship between the microbiome and host phenotype, on the read/sequence level. In this paper, we develop this novel deep learning approach and evaluate its application to amplicon sequences. We apply our approach to short DNA reads and full sequences of 16S ribosomal RNA (rRNA) marker genes, which identify the heterogeneity of a microbial community sample. We demonstrate that our implementation of a novel attention-based deep network architecture, Read2Pheno, achieves read-level phenotypic prediction. Training Read2Pheno models will encode sequences (reads) into dense, meaningful representations: learned embedded vectors output from the intermediate layer of the network model, which can provide biological insight when visualized. The attention layer of Read2Pheno models can also automatically identify nucleotide regions in reads/sequences which are particularly informative for classification. As such, this novel approach can avoid pre/post-processing and manual interpretation required with conventional approaches to microbiome sequence classification. We further show, as proof-of-concept, that aggregating read-level information can robustly predict microbial community properties, host phenotype, and taxonomic classification, with performance at least comparable to conventional approaches. An implementation of the attention-based deep learning network is available at https://github.com/EESI/sequence_attention (a python package) and https://github.com/EESI/seq2att (a command line tool). Public Library of Science 2021-09-22 /pmc/articles/PMC8496832/ /pubmed/34550967 http://dx.doi.org/10.1371/journal.pcbi.1009345 Text en © 2021 Zhao et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Zhao, Zhengqiao Woloszynek, Stephen Agbavor, Felix Mell, Joshua Chang Sokhansanj, Bahrad A. Rosen, Gail L. Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network |
title | Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network |
title_full | Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network |
title_fullStr | Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network |
title_full_unstemmed | Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network |
title_short | Learning, visualizing and exploring 16S rRNA structure using an attention-based deep neural network |
title_sort | learning, visualizing and exploring 16s rrna structure using an attention-based deep neural network |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8496832/ https://www.ncbi.nlm.nih.gov/pubmed/34550967 http://dx.doi.org/10.1371/journal.pcbi.1009345 |
work_keys_str_mv | AT zhaozhengqiao learningvisualizingandexploring16srrnastructureusinganattentionbaseddeepneuralnetwork AT woloszynekstephen learningvisualizingandexploring16srrnastructureusinganattentionbaseddeepneuralnetwork AT agbavorfelix learningvisualizingandexploring16srrnastructureusinganattentionbaseddeepneuralnetwork AT melljoshuachang learningvisualizingandexploring16srrnastructureusinganattentionbaseddeepneuralnetwork AT sokhansanjbahrada learningvisualizingandexploring16srrnastructureusinganattentionbaseddeepneuralnetwork AT rosengaill learningvisualizingandexploring16srrnastructureusinganattentionbaseddeepneuralnetwork |