Cargando…
Probabilistic generative transformer language models for generative design of molecules
Self-supervised neural language models have recently found wide applications in the generative design of organic molecules and protein sequences as well as representation learning for downstream structure classification and functional prediction. However, most of the existing deep learning models fo...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10518939/ https://www.ncbi.nlm.nih.gov/pubmed/37749655 http://dx.doi.org/10.1186/s13321-023-00759-z |
_version_ | 1785109620933525504 |
---|---|
author | Wei, Lai Fu, Nihang Song, Yuqi Wang, Qian Hu, Jianjun |
author_facet | Wei, Lai Fu, Nihang Song, Yuqi Wang, Qian Hu, Jianjun |
author_sort | Wei, Lai |
collection | PubMed |
description | Self-supervised neural language models have recently found wide applications in the generative design of organic molecules and protein sequences as well as representation learning for downstream structure classification and functional prediction. However, most of the existing deep learning models for molecule design usually require a big dataset and have a black-box architecture, which makes it difficult to interpret their design logic. Here we propose the Generative Molecular Transformer (GMTransformer), a probabilistic neural network model for generative design of molecules. Our model is built on the blank filling language model originally developed for text processing, which has demonstrated unique advantages in learning the “molecules grammars” with high-quality generation, interpretability, and data efficiency. Benchmarked on the MOSES datasets, our models achieve high novelty and Scaf compared to other baselines. The probabilistic generation steps have the potential in tinkering with molecule design due to their capability of recommending how to modify existing molecules with explanation, guided by the learned implicit molecule chemistry. The source code and datasets can be accessed freely at https://github.com/usccolumbia/GMTransformer |
format | Online Article Text |
id | pubmed-10518939 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-105189392023-09-26 Probabilistic generative transformer language models for generative design of molecules Wei, Lai Fu, Nihang Song, Yuqi Wang, Qian Hu, Jianjun J Cheminform Research Self-supervised neural language models have recently found wide applications in the generative design of organic molecules and protein sequences as well as representation learning for downstream structure classification and functional prediction. However, most of the existing deep learning models for molecule design usually require a big dataset and have a black-box architecture, which makes it difficult to interpret their design logic. Here we propose the Generative Molecular Transformer (GMTransformer), a probabilistic neural network model for generative design of molecules. Our model is built on the blank filling language model originally developed for text processing, which has demonstrated unique advantages in learning the “molecules grammars” with high-quality generation, interpretability, and data efficiency. Benchmarked on the MOSES datasets, our models achieve high novelty and Scaf compared to other baselines. The probabilistic generation steps have the potential in tinkering with molecule design due to their capability of recommending how to modify existing molecules with explanation, guided by the learned implicit molecule chemistry. The source code and datasets can be accessed freely at https://github.com/usccolumbia/GMTransformer Springer International Publishing 2023-09-25 /pmc/articles/PMC10518939/ /pubmed/37749655 http://dx.doi.org/10.1186/s13321-023-00759-z Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/ (https://creativecommons.org/publicdomain/zero/1.0/) ) applies to the data made available in this article, unless otherwise stated in a credit line to the data. |
spellingShingle | Research Wei, Lai Fu, Nihang Song, Yuqi Wang, Qian Hu, Jianjun Probabilistic generative transformer language models for generative design of molecules |
title | Probabilistic generative transformer language models for generative design of molecules |
title_full | Probabilistic generative transformer language models for generative design of molecules |
title_fullStr | Probabilistic generative transformer language models for generative design of molecules |
title_full_unstemmed | Probabilistic generative transformer language models for generative design of molecules |
title_short | Probabilistic generative transformer language models for generative design of molecules |
title_sort | probabilistic generative transformer language models for generative design of molecules |
topic | Research |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10518939/ https://www.ncbi.nlm.nih.gov/pubmed/37749655 http://dx.doi.org/10.1186/s13321-023-00759-z |
work_keys_str_mv | AT weilai probabilisticgenerativetransformerlanguagemodelsforgenerativedesignofmolecules AT funihang probabilisticgenerativetransformerlanguagemodelsforgenerativedesignofmolecules AT songyuqi probabilisticgenerativetransformerlanguagemodelsforgenerativedesignofmolecules AT wangqian probabilisticgenerativetransformerlanguagemodelsforgenerativedesignofmolecules AT hujianjun probabilisticgenerativetransformerlanguagemodelsforgenerativedesignofmolecules |