Cargando…
Attention-based generative models for de novo molecular design
Attention mechanisms have led to many breakthroughs in sequential data modeling but have yet to be incorporated into any generative algorithms for molecular design. Here we explore the impact of adding self-attention layers to generative β-VAE models and show that those with attention are able to le...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The Royal Society of Chemistry
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8221056/ https://www.ncbi.nlm.nih.gov/pubmed/34221317 http://dx.doi.org/10.1039/d1sc01050f |
_version_ | 1783711261185277952 |
---|---|
author | Dollar, Orion Joshi, Nisarg Beck, David A. C. Pfaendtner, Jim |
author_facet | Dollar, Orion Joshi, Nisarg Beck, David A. C. Pfaendtner, Jim |
author_sort | Dollar, Orion |
collection | PubMed |
description | Attention mechanisms have led to many breakthroughs in sequential data modeling but have yet to be incorporated into any generative algorithms for molecular design. Here we explore the impact of adding self-attention layers to generative β-VAE models and show that those with attention are able to learn a complex “molecular grammar” while improving performance on downstream tasks such as accurately sampling from the latent space (“model memory”) or exploring novel chemistries not present in the training data. There is a notable relationship between a model's architecture, the structure of its latent memory and its performance during inference. We demonstrate that there is an unavoidable tradeoff between model exploration and validity that is a function of the complexity of the latent memory. However, novel sampling schemes may be used that optimize this tradeoff. We anticipate that attention will play an important role in future molecular design algorithms that can make efficient use of the detailed molecular substructures learned by the transformer. |
format | Online Article Text |
id | pubmed-8221056 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | The Royal Society of Chemistry |
record_format | MEDLINE/PubMed |
spelling | pubmed-82210562021-07-02 Attention-based generative models for de novo molecular design Dollar, Orion Joshi, Nisarg Beck, David A. C. Pfaendtner, Jim Chem Sci Chemistry Attention mechanisms have led to many breakthroughs in sequential data modeling but have yet to be incorporated into any generative algorithms for molecular design. Here we explore the impact of adding self-attention layers to generative β-VAE models and show that those with attention are able to learn a complex “molecular grammar” while improving performance on downstream tasks such as accurately sampling from the latent space (“model memory”) or exploring novel chemistries not present in the training data. There is a notable relationship between a model's architecture, the structure of its latent memory and its performance during inference. We demonstrate that there is an unavoidable tradeoff between model exploration and validity that is a function of the complexity of the latent memory. However, novel sampling schemes may be used that optimize this tradeoff. We anticipate that attention will play an important role in future molecular design algorithms that can make efficient use of the detailed molecular substructures learned by the transformer. The Royal Society of Chemistry 2021-05-14 /pmc/articles/PMC8221056/ /pubmed/34221317 http://dx.doi.org/10.1039/d1sc01050f Text en This journal is © The Royal Society of Chemistry https://creativecommons.org/licenses/by-nc/3.0/ |
spellingShingle | Chemistry Dollar, Orion Joshi, Nisarg Beck, David A. C. Pfaendtner, Jim Attention-based generative models for de novo molecular design |
title | Attention-based generative models for de novo molecular design |
title_full | Attention-based generative models for de novo molecular design |
title_fullStr | Attention-based generative models for de novo molecular design |
title_full_unstemmed | Attention-based generative models for de novo molecular design |
title_short | Attention-based generative models for de novo molecular design |
title_sort | attention-based generative models for de novo molecular design |
topic | Chemistry |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8221056/ https://www.ncbi.nlm.nih.gov/pubmed/34221317 http://dx.doi.org/10.1039/d1sc01050f |
work_keys_str_mv | AT dollarorion attentionbasedgenerativemodelsfordenovomoleculardesign AT joshinisarg attentionbasedgenerativemodelsfordenovomoleculardesign AT beckdavidac attentionbasedgenerativemodelsfordenovomoleculardesign AT pfaendtnerjim attentionbasedgenerativemodelsfordenovomoleculardesign |