Cargando…

To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy

There is an increasing number of medical use cases where classification algorithms based on deep neural networks reach performance levels that are competitive with human medical experts. To alleviate the challenges of small dataset sizes, these systems often rely on pretraining. In this work, we aim...

Descripción completa

Detalles Bibliográficos
Autores principales: Srinivasan, Vignesh, Strodthoff, Nils, Ma, Jackie, Binder, Alexander, Müller, Klaus-Robert, Samek, Wojciech
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9578637/
https://www.ncbi.nlm.nih.gov/pubmed/36256665
http://dx.doi.org/10.1371/journal.pone.0274291
_version_ 1784812006748979200
author Srinivasan, Vignesh
Strodthoff, Nils
Ma, Jackie
Binder, Alexander
Müller, Klaus-Robert
Samek, Wojciech
author_facet Srinivasan, Vignesh
Strodthoff, Nils
Ma, Jackie
Binder, Alexander
Müller, Klaus-Robert
Samek, Wojciech
author_sort Srinivasan, Vignesh
collection PubMed
description There is an increasing number of medical use cases where classification algorithms based on deep neural networks reach performance levels that are competitive with human medical experts. To alleviate the challenges of small dataset sizes, these systems often rely on pretraining. In this work, we aim to assess the broader implications of these approaches in order to better understand what type of pretraining works reliably (with respect to performance, robustness, learned representation etc.) in practice and what type of pretraining dataset is best suited to achieve good performance in small target dataset size scenarios. Considering diabetic retinopathy grading as an exemplary use case, we compare the impact of different training procedures including recently established self-supervised pretraining methods based on contrastive learning. To this end, we investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions. Our results indicate that models initialized from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions. In particular, self-supervised models show further benefits to supervised models. Self-supervised models with initialization from ImageNet pretraining not only report higher performance, they also reduce overfitting to large lesions along with improvements in taking into account minute lesions indicative of the progression of the disease. Understanding the effects of pretraining in a broader sense that goes beyond simple performance comparisons is of crucial importance for the broader medical imaging community beyond the use case considered in this work.
format Online
Article
Text
id pubmed-9578637
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-95786372022-10-19 To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy Srinivasan, Vignesh Strodthoff, Nils Ma, Jackie Binder, Alexander Müller, Klaus-Robert Samek, Wojciech PLoS One Research Article There is an increasing number of medical use cases where classification algorithms based on deep neural networks reach performance levels that are competitive with human medical experts. To alleviate the challenges of small dataset sizes, these systems often rely on pretraining. In this work, we aim to assess the broader implications of these approaches in order to better understand what type of pretraining works reliably (with respect to performance, robustness, learned representation etc.) in practice and what type of pretraining dataset is best suited to achieve good performance in small target dataset size scenarios. Considering diabetic retinopathy grading as an exemplary use case, we compare the impact of different training procedures including recently established self-supervised pretraining methods based on contrastive learning. To this end, we investigate different aspects such as quantitative performance, statistics of the learned feature representations, interpretability and robustness to image distortions. Our results indicate that models initialized from ImageNet pretraining report a significant increase in performance, generalization and robustness to image distortions. In particular, self-supervised models show further benefits to supervised models. Self-supervised models with initialization from ImageNet pretraining not only report higher performance, they also reduce overfitting to large lesions along with improvements in taking into account minute lesions indicative of the progression of the disease. Understanding the effects of pretraining in a broader sense that goes beyond simple performance comparisons is of crucial importance for the broader medical imaging community beyond the use case considered in this work. Public Library of Science 2022-10-18 /pmc/articles/PMC9578637/ /pubmed/36256665 http://dx.doi.org/10.1371/journal.pone.0274291 Text en © 2022 Srinivasan et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Srinivasan, Vignesh
Strodthoff, Nils
Ma, Jackie
Binder, Alexander
Müller, Klaus-Robert
Samek, Wojciech
To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy
title To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy
title_full To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy
title_fullStr To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy
title_full_unstemmed To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy
title_short To pretrain or not? A systematic analysis of the benefits of pretraining in diabetic retinopathy
title_sort to pretrain or not? a systematic analysis of the benefits of pretraining in diabetic retinopathy
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9578637/
https://www.ncbi.nlm.nih.gov/pubmed/36256665
http://dx.doi.org/10.1371/journal.pone.0274291
work_keys_str_mv AT srinivasanvignesh topretrainornotasystematicanalysisofthebenefitsofpretrainingindiabeticretinopathy
AT strodthoffnils topretrainornotasystematicanalysisofthebenefitsofpretrainingindiabeticretinopathy
AT majackie topretrainornotasystematicanalysisofthebenefitsofpretrainingindiabeticretinopathy
AT binderalexander topretrainornotasystematicanalysisofthebenefitsofpretrainingindiabeticretinopathy
AT mullerklausrobert topretrainornotasystematicanalysisofthebenefitsofpretrainingindiabeticretinopathy
AT samekwojciech topretrainornotasystematicanalysisofthebenefitsofpretrainingindiabeticretinopathy