Cargando…
Meta-learning for transformer-based prediction of potent compounds
For many machine learning applications in drug discovery, only limited amounts of training data are available. This typically applies to compound design and activity prediction and often restricts machine learning, especially deep learning. For low-data applications, specialized learning strategies...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10522638/ https://www.ncbi.nlm.nih.gov/pubmed/37752164 http://dx.doi.org/10.1038/s41598-023-43046-5 |
_version_ | 1785110394301317120 |
---|---|
author | Chen, Hengwei Bajorath, Jürgen |
author_facet | Chen, Hengwei Bajorath, Jürgen |
author_sort | Chen, Hengwei |
collection | PubMed |
description | For many machine learning applications in drug discovery, only limited amounts of training data are available. This typically applies to compound design and activity prediction and often restricts machine learning, especially deep learning. For low-data applications, specialized learning strategies can be considered to limit required training data. Among these is meta-learning that attempts to enable learning in low-data regimes by combining outputs of different models and utilizing meta-data from these predictions. However, in drug discovery settings, meta-learning is still in its infancy. In this study, we have explored meta-learning for the prediction of potent compounds via generative design using transformer models. For different activity classes, meta-learning models were derived to predict highly potent compounds from weakly potent templates in the presence of varying amounts of fine-tuning data and compared to other transformers developed for this task. Meta-learning consistently led to statistically significant improvements in model performance, in particular, when fine-tuning data were limited. Moreover, meta-learning models generated target compounds with higher potency and larger potency differences between templates and targets than other transformers, indicating their potential for low-data compound design. |
format | Online Article Text |
id | pubmed-10522638 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-105226382023-09-28 Meta-learning for transformer-based prediction of potent compounds Chen, Hengwei Bajorath, Jürgen Sci Rep Article For many machine learning applications in drug discovery, only limited amounts of training data are available. This typically applies to compound design and activity prediction and often restricts machine learning, especially deep learning. For low-data applications, specialized learning strategies can be considered to limit required training data. Among these is meta-learning that attempts to enable learning in low-data regimes by combining outputs of different models and utilizing meta-data from these predictions. However, in drug discovery settings, meta-learning is still in its infancy. In this study, we have explored meta-learning for the prediction of potent compounds via generative design using transformer models. For different activity classes, meta-learning models were derived to predict highly potent compounds from weakly potent templates in the presence of varying amounts of fine-tuning data and compared to other transformers developed for this task. Meta-learning consistently led to statistically significant improvements in model performance, in particular, when fine-tuning data were limited. Moreover, meta-learning models generated target compounds with higher potency and larger potency differences between templates and targets than other transformers, indicating their potential for low-data compound design. Nature Publishing Group UK 2023-09-26 /pmc/articles/PMC10522638/ /pubmed/37752164 http://dx.doi.org/10.1038/s41598-023-43046-5 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Chen, Hengwei Bajorath, Jürgen Meta-learning for transformer-based prediction of potent compounds |
title | Meta-learning for transformer-based prediction of potent compounds |
title_full | Meta-learning for transformer-based prediction of potent compounds |
title_fullStr | Meta-learning for transformer-based prediction of potent compounds |
title_full_unstemmed | Meta-learning for transformer-based prediction of potent compounds |
title_short | Meta-learning for transformer-based prediction of potent compounds |
title_sort | meta-learning for transformer-based prediction of potent compounds |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10522638/ https://www.ncbi.nlm.nih.gov/pubmed/37752164 http://dx.doi.org/10.1038/s41598-023-43046-5 |
work_keys_str_mv | AT chenhengwei metalearningfortransformerbasedpredictionofpotentcompounds AT bajorathjurgen metalearningfortransformerbasedpredictionofpotentcompounds |