Cargando…
Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimic...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8575004/ https://www.ncbi.nlm.nih.gov/pubmed/34169324 http://dx.doi.org/10.1093/bib/bbab225 |
_version_ | 1784595600911630336 |
---|---|
author | Li, Hongjian Lu, Gang Sze, Kam-Heung Su, Xianwei Chan, Wai-Yee Leung, Kwong-Sak |
author_facet | Li, Hongjian Lu, Gang Sze, Kam-Heung Su, Xianwei Chan, Wai-Yee Leung, Kwong-Sak |
author_sort | Li, Hongjian |
collection | PubMed |
description | The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimicking the process of prospective prediction of binding affinity, we have evaluated three broadly used classical scoring functions and five machine-learning counterparts calibrated with both random forest and extreme gradient boosting using both solo and hybrid features, showing for the first time that machine-learning scoring functions trained exclusively on a proportion of as low as 8% complexes dissimilar to the test set already outperform classical scoring functions, a percentage that is far lower than what has been recently reported on all the three CASF benchmarks. The performance of machine-learning scoring functions is underestimated due to the absence of similar samples in some artificially created training sets that discard the full spectrum of complexes to be found in a prospective environment. Given the inevitability of any degree of similarity contained in a large dataset, the criteria for scoring function selection depend on which one can make the best use of all available materials. Software code and data are provided at https://github.com/cusdulab/MLSF for interested readers to rapidly rebuild the scoring functions and reproduce our results, even to make extended analyses on their own benchmarks. |
format | Online Article Text |
id | pubmed-8575004 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-85750042021-11-09 Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark Li, Hongjian Lu, Gang Sze, Kam-Heung Su, Xianwei Chan, Wai-Yee Leung, Kwong-Sak Brief Bioinform Problem Solving Protocol The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimicking the process of prospective prediction of binding affinity, we have evaluated three broadly used classical scoring functions and five machine-learning counterparts calibrated with both random forest and extreme gradient boosting using both solo and hybrid features, showing for the first time that machine-learning scoring functions trained exclusively on a proportion of as low as 8% complexes dissimilar to the test set already outperform classical scoring functions, a percentage that is far lower than what has been recently reported on all the three CASF benchmarks. The performance of machine-learning scoring functions is underestimated due to the absence of similar samples in some artificially created training sets that discard the full spectrum of complexes to be found in a prospective environment. Given the inevitability of any degree of similarity contained in a large dataset, the criteria for scoring function selection depend on which one can make the best use of all available materials. Software code and data are provided at https://github.com/cusdulab/MLSF for interested readers to rapidly rebuild the scoring functions and reproduce our results, even to make extended analyses on their own benchmarks. Oxford University Press 2021-06-24 /pmc/articles/PMC8575004/ /pubmed/34169324 http://dx.doi.org/10.1093/bib/bbab225 Text en © The Author(s) 2021. Published by Oxford University Press. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Problem Solving Protocol Li, Hongjian Lu, Gang Sze, Kam-Heung Su, Xianwei Chan, Wai-Yee Leung, Kwong-Sak Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
title | Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
title_full | Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
title_fullStr | Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
title_full_unstemmed | Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
title_short | Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
title_sort | machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark |
topic | Problem Solving Protocol |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8575004/ https://www.ncbi.nlm.nih.gov/pubmed/34169324 http://dx.doi.org/10.1093/bib/bbab225 |
work_keys_str_mv | AT lihongjian machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark AT lugang machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark AT szekamheung machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark AT suxianwei machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark AT chanwaiyee machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark AT leungkwongsak machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark |