Cargando…

Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark

The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimic...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Hongjian, Lu, Gang, Sze, Kam-Heung, Su, Xianwei, Chan, Wai-Yee, Leung, Kwong-Sak
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8575004/
https://www.ncbi.nlm.nih.gov/pubmed/34169324
http://dx.doi.org/10.1093/bib/bbab225
_version_ 1784595600911630336
author Li, Hongjian
Lu, Gang
Sze, Kam-Heung
Su, Xianwei
Chan, Wai-Yee
Leung, Kwong-Sak
author_facet Li, Hongjian
Lu, Gang
Sze, Kam-Heung
Su, Xianwei
Chan, Wai-Yee
Leung, Kwong-Sak
author_sort Li, Hongjian
collection PubMed
description The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimicking the process of prospective prediction of binding affinity, we have evaluated three broadly used classical scoring functions and five machine-learning counterparts calibrated with both random forest and extreme gradient boosting using both solo and hybrid features, showing for the first time that machine-learning scoring functions trained exclusively on a proportion of as low as 8% complexes dissimilar to the test set already outperform classical scoring functions, a percentage that is far lower than what has been recently reported on all the three CASF benchmarks. The performance of machine-learning scoring functions is underestimated due to the absence of similar samples in some artificially created training sets that discard the full spectrum of complexes to be found in a prospective environment. Given the inevitability of any degree of similarity contained in a large dataset, the criteria for scoring function selection depend on which one can make the best use of all available materials. Software code and data are provided at https://github.com/cusdulab/MLSF for interested readers to rapidly rebuild the scoring functions and reproduce our results, even to make extended analyses on their own benchmarks.
format Online
Article
Text
id pubmed-8575004
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-85750042021-11-09 Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark Li, Hongjian Lu, Gang Sze, Kam-Heung Su, Xianwei Chan, Wai-Yee Leung, Kwong-Sak Brief Bioinform Problem Solving Protocol The superior performance of machine-learning scoring functions for docking has caused a series of debates on whether it is due to learning knowledge from training data that are similar in some sense to the test data. With a systematically revised methodology and a blind benchmark realistically mimicking the process of prospective prediction of binding affinity, we have evaluated three broadly used classical scoring functions and five machine-learning counterparts calibrated with both random forest and extreme gradient boosting using both solo and hybrid features, showing for the first time that machine-learning scoring functions trained exclusively on a proportion of as low as 8% complexes dissimilar to the test set already outperform classical scoring functions, a percentage that is far lower than what has been recently reported on all the three CASF benchmarks. The performance of machine-learning scoring functions is underestimated due to the absence of similar samples in some artificially created training sets that discard the full spectrum of complexes to be found in a prospective environment. Given the inevitability of any degree of similarity contained in a large dataset, the criteria for scoring function selection depend on which one can make the best use of all available materials. Software code and data are provided at https://github.com/cusdulab/MLSF for interested readers to rapidly rebuild the scoring functions and reproduce our results, even to make extended analyses on their own benchmarks. Oxford University Press 2021-06-24 /pmc/articles/PMC8575004/ /pubmed/34169324 http://dx.doi.org/10.1093/bib/bbab225 Text en © The Author(s) 2021. Published by Oxford University Press. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Problem Solving Protocol
Li, Hongjian
Lu, Gang
Sze, Kam-Heung
Su, Xianwei
Chan, Wai-Yee
Leung, Kwong-Sak
Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
title Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
title_full Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
title_fullStr Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
title_full_unstemmed Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
title_short Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
title_sort machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark
topic Problem Solving Protocol
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8575004/
https://www.ncbi.nlm.nih.gov/pubmed/34169324
http://dx.doi.org/10.1093/bib/bbab225
work_keys_str_mv AT lihongjian machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark
AT lugang machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark
AT szekamheung machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark
AT suxianwei machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark
AT chanwaiyee machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark
AT leungkwongsak machinelearningscoringfunctionstrainedoncomplexesdissimilartothetestsetalreadyoutperformclassicalcounterpartsonablindbenchmark