Cargando…
Reproducibility of deep learning in digital pathology whole slide image analysis
For a method to be widely adopted in medical research or clinical practice, it needs to be reproducible so that clinicians and regulators can have confidence in its use. Machine learning and deep learning have a particular set of challenges around reproducibility. Small differences in the settings o...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931349/ https://www.ncbi.nlm.nih.gov/pubmed/36812609 http://dx.doi.org/10.1371/journal.pdig.0000145 |
_version_ | 1784889229750304768 |
---|---|
author | Fell, Christina Mohammadi, Mahnaz Morrison, David Arandjelovic, Ognjen Caie, Peter Harris-Birtill, David |
author_facet | Fell, Christina Mohammadi, Mahnaz Morrison, David Arandjelovic, Ognjen Caie, Peter Harris-Birtill, David |
author_sort | Fell, Christina |
collection | PubMed |
description | For a method to be widely adopted in medical research or clinical practice, it needs to be reproducible so that clinicians and regulators can have confidence in its use. Machine learning and deep learning have a particular set of challenges around reproducibility. Small differences in the settings or the data used for training a model can lead to large differences in the outcomes of experiments. In this work, three top-performing algorithms from the Camelyon grand challenges are reproduced using only information presented in the associated papers and the results are then compared to those reported. Seemingly minor details were found to be critical to performance and yet their importance is difficult to appreciate until the actual reproduction is attempted. We observed that authors generally describe the key technical aspects of their models well but fail to maintain the same reporting standards when it comes to data preprocessing which is essential to reproducibility. As an important contribution of the present study and its findings, we introduce a reproducibility checklist that tabulates information that needs to be reported in histopathology ML-based work in order to make it reproducible. |
format | Online Article Text |
id | pubmed-9931349 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-99313492023-02-16 Reproducibility of deep learning in digital pathology whole slide image analysis Fell, Christina Mohammadi, Mahnaz Morrison, David Arandjelovic, Ognjen Caie, Peter Harris-Birtill, David PLOS Digit Health Research Article For a method to be widely adopted in medical research or clinical practice, it needs to be reproducible so that clinicians and regulators can have confidence in its use. Machine learning and deep learning have a particular set of challenges around reproducibility. Small differences in the settings or the data used for training a model can lead to large differences in the outcomes of experiments. In this work, three top-performing algorithms from the Camelyon grand challenges are reproduced using only information presented in the associated papers and the results are then compared to those reported. Seemingly minor details were found to be critical to performance and yet their importance is difficult to appreciate until the actual reproduction is attempted. We observed that authors generally describe the key technical aspects of their models well but fail to maintain the same reporting standards when it comes to data preprocessing which is essential to reproducibility. As an important contribution of the present study and its findings, we introduce a reproducibility checklist that tabulates information that needs to be reported in histopathology ML-based work in order to make it reproducible. Public Library of Science 2022-12-02 /pmc/articles/PMC9931349/ /pubmed/36812609 http://dx.doi.org/10.1371/journal.pdig.0000145 Text en © 2022 Fell et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Fell, Christina Mohammadi, Mahnaz Morrison, David Arandjelovic, Ognjen Caie, Peter Harris-Birtill, David Reproducibility of deep learning in digital pathology whole slide image analysis |
title | Reproducibility of deep learning in digital pathology whole slide image analysis |
title_full | Reproducibility of deep learning in digital pathology whole slide image analysis |
title_fullStr | Reproducibility of deep learning in digital pathology whole slide image analysis |
title_full_unstemmed | Reproducibility of deep learning in digital pathology whole slide image analysis |
title_short | Reproducibility of deep learning in digital pathology whole slide image analysis |
title_sort | reproducibility of deep learning in digital pathology whole slide image analysis |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931349/ https://www.ncbi.nlm.nih.gov/pubmed/36812609 http://dx.doi.org/10.1371/journal.pdig.0000145 |
work_keys_str_mv | AT fellchristina reproducibilityofdeeplearningindigitalpathologywholeslideimageanalysis AT mohammadimahnaz reproducibilityofdeeplearningindigitalpathologywholeslideimageanalysis AT morrisondavid reproducibilityofdeeplearningindigitalpathologywholeslideimageanalysis AT arandjelovicognjen reproducibilityofdeeplearningindigitalpathologywholeslideimageanalysis AT caiepeter reproducibilityofdeeplearningindigitalpathologywholeslideimageanalysis AT harrisbirtilldavid reproducibilityofdeeplearningindigitalpathologywholeslideimageanalysis |