Cargando…

Systematic misestimation of machine learning performance in neuroimaging studies of depression

We currently observe a disconcerting phenomenon in machine learning studies in psychiatry: While we would expect larger samples to yield better results due to the availability of more data, larger machine learning studies consistently show much weaker performance than the numerous small-scale studie...

Descripción completa

Detalles Bibliográficos
Autores principales: Flint, Claas, Cearns, Micah, Opel, Nils, Redlich, Ronny, Mehler, David M. A., Emden, Daniel, Winter, Nils R., Leenings, Ramona, Eickhoff, Simon B., Kircher, Tilo, Krug, Axel, Nenadic, Igor, Arolt, Volker, Clark, Scott, Baune, Bernhard T., Jiang, Xiaoyi, Dannlowski, Udo, Hahn, Tim
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer International Publishing 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8209109/
https://www.ncbi.nlm.nih.gov/pubmed/33958703
http://dx.doi.org/10.1038/s41386-021-01020-7
Descripción
Sumario:We currently observe a disconcerting phenomenon in machine learning studies in psychiatry: While we would expect larger samples to yield better results due to the availability of more data, larger machine learning studies consistently show much weaker performance than the numerous small-scale studies. Here, we systematically investigated this effect focusing on one of the most heavily studied questions in the field, namely the classification of patients suffering from Major Depressive Disorder (MDD) and healthy controls based on neuroimaging data. Drawing upon structural MRI data from a balanced sample of N = 1868 MDD patients and healthy controls from our recent international Predictive Analytics Competition (PAC), we first trained and tested a classification model on the full dataset which yielded an accuracy of 61%. Next, we mimicked the process by which researchers would draw samples of various sizes (N = 4 to N = 150) from the population and showed a strong risk of misestimation. Specifically, for small sample sizes (N = 20), we observe accuracies of up to 95%. For medium sample sizes (N = 100) accuracies up to 75% were found. Importantly, further investigation showed that sufficiently large test sets effectively protect against performance misestimation whereas larger datasets per se do not. While these results question the validity of a substantial part of the current literature, we outline the relatively low-cost remedy of larger test sets, which is readily available in most cases.