Cargando…

Attention-based generative models for de novo molecular design

Attention mechanisms have led to many breakthroughs in sequential data modeling but have yet to be incorporated into any generative algorithms for molecular design. Here we explore the impact of adding self-attention layers to generative β-VAE models and show that those with attention are able to le...

Descripción completa

Detalles Bibliográficos
Autores principales: Dollar, Orion, Joshi, Nisarg, Beck, David A. C., Pfaendtner, Jim
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Royal Society of Chemistry 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8221056/
https://www.ncbi.nlm.nih.gov/pubmed/34221317
http://dx.doi.org/10.1039/d1sc01050f
_version_ 1783711261185277952
author Dollar, Orion
Joshi, Nisarg
Beck, David A. C.
Pfaendtner, Jim
author_facet Dollar, Orion
Joshi, Nisarg
Beck, David A. C.
Pfaendtner, Jim
author_sort Dollar, Orion
collection PubMed
description Attention mechanisms have led to many breakthroughs in sequential data modeling but have yet to be incorporated into any generative algorithms for molecular design. Here we explore the impact of adding self-attention layers to generative β-VAE models and show that those with attention are able to learn a complex “molecular grammar” while improving performance on downstream tasks such as accurately sampling from the latent space (“model memory”) or exploring novel chemistries not present in the training data. There is a notable relationship between a model's architecture, the structure of its latent memory and its performance during inference. We demonstrate that there is an unavoidable tradeoff between model exploration and validity that is a function of the complexity of the latent memory. However, novel sampling schemes may be used that optimize this tradeoff. We anticipate that attention will play an important role in future molecular design algorithms that can make efficient use of the detailed molecular substructures learned by the transformer.
format Online
Article
Text
id pubmed-8221056
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher The Royal Society of Chemistry
record_format MEDLINE/PubMed
spelling pubmed-82210562021-07-02 Attention-based generative models for de novo molecular design Dollar, Orion Joshi, Nisarg Beck, David A. C. Pfaendtner, Jim Chem Sci Chemistry Attention mechanisms have led to many breakthroughs in sequential data modeling but have yet to be incorporated into any generative algorithms for molecular design. Here we explore the impact of adding self-attention layers to generative β-VAE models and show that those with attention are able to learn a complex “molecular grammar” while improving performance on downstream tasks such as accurately sampling from the latent space (“model memory”) or exploring novel chemistries not present in the training data. There is a notable relationship between a model's architecture, the structure of its latent memory and its performance during inference. We demonstrate that there is an unavoidable tradeoff between model exploration and validity that is a function of the complexity of the latent memory. However, novel sampling schemes may be used that optimize this tradeoff. We anticipate that attention will play an important role in future molecular design algorithms that can make efficient use of the detailed molecular substructures learned by the transformer. The Royal Society of Chemistry 2021-05-14 /pmc/articles/PMC8221056/ /pubmed/34221317 http://dx.doi.org/10.1039/d1sc01050f Text en This journal is © The Royal Society of Chemistry https://creativecommons.org/licenses/by-nc/3.0/
spellingShingle Chemistry
Dollar, Orion
Joshi, Nisarg
Beck, David A. C.
Pfaendtner, Jim
Attention-based generative models for de novo molecular design
title Attention-based generative models for de novo molecular design
title_full Attention-based generative models for de novo molecular design
title_fullStr Attention-based generative models for de novo molecular design
title_full_unstemmed Attention-based generative models for de novo molecular design
title_short Attention-based generative models for de novo molecular design
title_sort attention-based generative models for de novo molecular design
topic Chemistry
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8221056/
https://www.ncbi.nlm.nih.gov/pubmed/34221317
http://dx.doi.org/10.1039/d1sc01050f
work_keys_str_mv AT dollarorion attentionbasedgenerativemodelsfordenovomoleculardesign
AT joshinisarg attentionbasedgenerativemodelsfordenovomoleculardesign
AT beckdavidac attentionbasedgenerativemodelsfordenovomoleculardesign
AT pfaendtnerjim attentionbasedgenerativemodelsfordenovomoleculardesign