Cargando…
Transformer-CNN: Swiss knife for QSAR modeling and interpretation
We present SMILES-embeddings derived from the internal encoder state of a Transformer [1] model trained to canonize SMILES as a Seq2Seq problem. Using a CharNN [2] architecture upon the embeddings results in higher quality interpretable QSAR/QSPR models on diverse benchmark datasets including regres...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7079452/ https://www.ncbi.nlm.nih.gov/pubmed/33431004 http://dx.doi.org/10.1186/s13321-020-00423-w |
_version_ | 1783507826940837888 |
---|---|
author | Karpov, Pavel Godin, Guillaume Tetko, Igor V. |
author_facet | Karpov, Pavel Godin, Guillaume Tetko, Igor V. |
author_sort | Karpov, Pavel |
collection | PubMed |
description | We present SMILES-embeddings derived from the internal encoder state of a Transformer [1] model trained to canonize SMILES as a Seq2Seq problem. Using a CharNN [2] architecture upon the embeddings results in higher quality interpretable QSAR/QSPR models on diverse benchmark datasets including regression and classification tasks. The proposed Transformer-CNN method uses SMILES augmentation for training and inference, and thus the prognosis is based on an internal consensus. That both the augmentation and transfer learning are based on embeddings allows the method to provide good results for small datasets. We discuss the reasons for such effectiveness and draft future directions for the development of the method. The source code and the embeddings needed to train a QSAR model are available on https://github.com/bigchem/transformer-cnn. The repository also has a standalone program for QSAR prognosis which calculates individual atoms contributions, thus interpreting the model’s result. OCHEM [3] environment (https://ochem.eu) hosts the on-line implementation of the method proposed. |
format | Online Article Text |
id | pubmed-7079452 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-70794522020-03-23 Transformer-CNN: Swiss knife for QSAR modeling and interpretation Karpov, Pavel Godin, Guillaume Tetko, Igor V. J Cheminform Research Article We present SMILES-embeddings derived from the internal encoder state of a Transformer [1] model trained to canonize SMILES as a Seq2Seq problem. Using a CharNN [2] architecture upon the embeddings results in higher quality interpretable QSAR/QSPR models on diverse benchmark datasets including regression and classification tasks. The proposed Transformer-CNN method uses SMILES augmentation for training and inference, and thus the prognosis is based on an internal consensus. That both the augmentation and transfer learning are based on embeddings allows the method to provide good results for small datasets. We discuss the reasons for such effectiveness and draft future directions for the development of the method. The source code and the embeddings needed to train a QSAR model are available on https://github.com/bigchem/transformer-cnn. The repository also has a standalone program for QSAR prognosis which calculates individual atoms contributions, thus interpreting the model’s result. OCHEM [3] environment (https://ochem.eu) hosts the on-line implementation of the method proposed. Springer International Publishing 2020-03-18 /pmc/articles/PMC7079452/ /pubmed/33431004 http://dx.doi.org/10.1186/s13321-020-00423-w Text en © The Author(s) 2020 Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data. |
spellingShingle | Research Article Karpov, Pavel Godin, Guillaume Tetko, Igor V. Transformer-CNN: Swiss knife for QSAR modeling and interpretation |
title | Transformer-CNN: Swiss knife for QSAR modeling and interpretation |
title_full | Transformer-CNN: Swiss knife for QSAR modeling and interpretation |
title_fullStr | Transformer-CNN: Swiss knife for QSAR modeling and interpretation |
title_full_unstemmed | Transformer-CNN: Swiss knife for QSAR modeling and interpretation |
title_short | Transformer-CNN: Swiss knife for QSAR modeling and interpretation |
title_sort | transformer-cnn: swiss knife for qsar modeling and interpretation |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7079452/ https://www.ncbi.nlm.nih.gov/pubmed/33431004 http://dx.doi.org/10.1186/s13321-020-00423-w |
work_keys_str_mv | AT karpovpavel transformercnnswissknifeforqsarmodelingandinterpretation AT godinguillaume transformercnnswissknifeforqsarmodelingandinterpretation AT tetkoigorv transformercnnswissknifeforqsarmodelingandinterpretation |