Cargando…
Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators
It is well-established that neural networks can predict or identify structural motifs of non-coding RNAs (ncRNAs). Yet, the neural network based identification of RNA structural motifs is limited by the availability of training data that are often insufficient for learning features of specific ncRNA...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9262186/ https://www.ncbi.nlm.nih.gov/pubmed/35797361 http://dx.doi.org/10.1371/journal.pcbi.1010240 |
_version_ | 1784742435496132608 |
---|---|
author | Brandenburg, Vivian B. Narberhaus, Franz Mosig, Axel |
author_facet | Brandenburg, Vivian B. Narberhaus, Franz Mosig, Axel |
author_sort | Brandenburg, Vivian B. |
collection | PubMed |
description | It is well-established that neural networks can predict or identify structural motifs of non-coding RNAs (ncRNAs). Yet, the neural network based identification of RNA structural motifs is limited by the availability of training data that are often insufficient for learning features of specific ncRNA families or structural motifs. Aiming to reliably identify intrinsic transcription terminators in bacteria, we introduce a novel pre-training approach that uses inverse folding to generate training data for predicting or identifying a specific family or structural motif of ncRNA. We assess the ability of neural networks to identify secondary structure by systematic in silico mutagenesis experiments. In a study to identify intrinsic transcription terminators as functionally well-understood RNA structural motifs, our inverse folding based pre-training approach significantly boosts the performance of neural network topologies, which outperform previous approaches to identify intrinsic transcription terminators. Inverse-folding based pre-training provides a simple, yet highly effective way to integrate the well-established thermodynamic energy model into deep neural networks for identifying ncRNA families or motifs. The pre-training technique is broadly applicable to a range of network topologies as well as different types of ncRNA families and motifs. |
format | Online Article Text |
id | pubmed-9262186 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-92621862022-07-08 Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators Brandenburg, Vivian B. Narberhaus, Franz Mosig, Axel PLoS Comput Biol Research Article It is well-established that neural networks can predict or identify structural motifs of non-coding RNAs (ncRNAs). Yet, the neural network based identification of RNA structural motifs is limited by the availability of training data that are often insufficient for learning features of specific ncRNA families or structural motifs. Aiming to reliably identify intrinsic transcription terminators in bacteria, we introduce a novel pre-training approach that uses inverse folding to generate training data for predicting or identifying a specific family or structural motif of ncRNA. We assess the ability of neural networks to identify secondary structure by systematic in silico mutagenesis experiments. In a study to identify intrinsic transcription terminators as functionally well-understood RNA structural motifs, our inverse folding based pre-training approach significantly boosts the performance of neural network topologies, which outperform previous approaches to identify intrinsic transcription terminators. Inverse-folding based pre-training provides a simple, yet highly effective way to integrate the well-established thermodynamic energy model into deep neural networks for identifying ncRNA families or motifs. The pre-training technique is broadly applicable to a range of network topologies as well as different types of ncRNA families and motifs. Public Library of Science 2022-07-07 /pmc/articles/PMC9262186/ /pubmed/35797361 http://dx.doi.org/10.1371/journal.pcbi.1010240 Text en © 2022 Brandenburg et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Brandenburg, Vivian B. Narberhaus, Franz Mosig, Axel Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
title | Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
title_full | Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
title_fullStr | Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
title_full_unstemmed | Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
title_short | Inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
title_sort | inverse folding based pre-training for the reliable identification of intrinsic transcription terminators |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9262186/ https://www.ncbi.nlm.nih.gov/pubmed/35797361 http://dx.doi.org/10.1371/journal.pcbi.1010240 |
work_keys_str_mv | AT brandenburgvivianb inversefoldingbasedpretrainingforthereliableidentificationofintrinsictranscriptionterminators AT narberhausfranz inversefoldingbasedpretrainingforthereliableidentificationofintrinsictranscriptionterminators AT mosigaxel inversefoldingbasedpretrainingforthereliableidentificationofintrinsictranscriptionterminators |