Cargando…

Meta-learning for transformer-based prediction of potent compounds

For many machine learning applications in drug discovery, only limited amounts of training data are available. This typically applies to compound design and activity prediction and often restricts machine learning, especially deep learning. For low-data applications, specialized learning strategies...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Hengwei, Bajorath, Jürgen
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10522638/
https://www.ncbi.nlm.nih.gov/pubmed/37752164
http://dx.doi.org/10.1038/s41598-023-43046-5
_version_ 1785110394301317120
author Chen, Hengwei
Bajorath, Jürgen
author_facet Chen, Hengwei
Bajorath, Jürgen
author_sort Chen, Hengwei
collection PubMed
description For many machine learning applications in drug discovery, only limited amounts of training data are available. This typically applies to compound design and activity prediction and often restricts machine learning, especially deep learning. For low-data applications, specialized learning strategies can be considered to limit required training data. Among these is meta-learning that attempts to enable learning in low-data regimes by combining outputs of different models and utilizing meta-data from these predictions. However, in drug discovery settings, meta-learning is still in its infancy. In this study, we have explored meta-learning for the prediction of potent compounds via generative design using transformer models. For different activity classes, meta-learning models were derived to predict highly potent compounds from weakly potent templates in the presence of varying amounts of fine-tuning data and compared to other transformers developed for this task. Meta-learning consistently led to statistically significant improvements in model performance, in particular, when fine-tuning data were limited. Moreover, meta-learning models generated target compounds with higher potency and larger potency differences between templates and targets than other transformers, indicating their potential for low-data compound design.
format Online
Article
Text
id pubmed-10522638
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-105226382023-09-28 Meta-learning for transformer-based prediction of potent compounds Chen, Hengwei Bajorath, Jürgen Sci Rep Article For many machine learning applications in drug discovery, only limited amounts of training data are available. This typically applies to compound design and activity prediction and often restricts machine learning, especially deep learning. For low-data applications, specialized learning strategies can be considered to limit required training data. Among these is meta-learning that attempts to enable learning in low-data regimes by combining outputs of different models and utilizing meta-data from these predictions. However, in drug discovery settings, meta-learning is still in its infancy. In this study, we have explored meta-learning for the prediction of potent compounds via generative design using transformer models. For different activity classes, meta-learning models were derived to predict highly potent compounds from weakly potent templates in the presence of varying amounts of fine-tuning data and compared to other transformers developed for this task. Meta-learning consistently led to statistically significant improvements in model performance, in particular, when fine-tuning data were limited. Moreover, meta-learning models generated target compounds with higher potency and larger potency differences between templates and targets than other transformers, indicating their potential for low-data compound design. Nature Publishing Group UK 2023-09-26 /pmc/articles/PMC10522638/ /pubmed/37752164 http://dx.doi.org/10.1038/s41598-023-43046-5 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Chen, Hengwei
Bajorath, Jürgen
Meta-learning for transformer-based prediction of potent compounds
title Meta-learning for transformer-based prediction of potent compounds
title_full Meta-learning for transformer-based prediction of potent compounds
title_fullStr Meta-learning for transformer-based prediction of potent compounds
title_full_unstemmed Meta-learning for transformer-based prediction of potent compounds
title_short Meta-learning for transformer-based prediction of potent compounds
title_sort meta-learning for transformer-based prediction of potent compounds
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10522638/
https://www.ncbi.nlm.nih.gov/pubmed/37752164
http://dx.doi.org/10.1038/s41598-023-43046-5
work_keys_str_mv AT chenhengwei metalearningfortransformerbasedpredictionofpotentcompounds
AT bajorathjurgen metalearningfortransformerbasedpredictionofpotentcompounds