Cargando…

A divide and conquer approach to maximise deep learning mammography classification accuracies

Breast cancer claims 11,400 lives on average every year in the UK, making it one of the deadliest diseases. Mammography is the gold standard for detecting early signs of breast cancer, which can help cure the disease during its early stages. However, incorrect mammography diagnoses are common and ma...

Descripción completa

Detalles Bibliográficos
Autores principales: Jaamour, Adam, Myles, Craig, Patel, Ashay, Chen, Shuen-Jen, McMillan, Lewis, Harris-Birtill, David
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10218746/
https://www.ncbi.nlm.nih.gov/pubmed/37235566
http://dx.doi.org/10.1371/journal.pone.0280841
_version_ 1785048846993195008
author Jaamour, Adam
Myles, Craig
Patel, Ashay
Chen, Shuen-Jen
McMillan, Lewis
Harris-Birtill, David
author_facet Jaamour, Adam
Myles, Craig
Patel, Ashay
Chen, Shuen-Jen
McMillan, Lewis
Harris-Birtill, David
author_sort Jaamour, Adam
collection PubMed
description Breast cancer claims 11,400 lives on average every year in the UK, making it one of the deadliest diseases. Mammography is the gold standard for detecting early signs of breast cancer, which can help cure the disease during its early stages. However, incorrect mammography diagnoses are common and may harm patients through unnecessary treatments and operations (or a lack of treatment). Therefore, systems that can learn to detect breast cancer on their own could help reduce the number of incorrect interpretations and missed cases. Various deep learning techniques, which can be used to implement a system that learns how to detect instances of breast cancer in mammograms, are explored throughout this paper. Convolution Neural Networks (CNNs) are used as part of a pipeline based on deep learning techniques. A divide and conquer approach is followed to analyse the effects on performance and efficiency when utilising diverse deep learning techniques such as varying network architectures (VGG19, ResNet50, InceptionV3, DenseNet121, MobileNetV2), class weights, input sizes, image ratios, pre-processing techniques, transfer learning, dropout rates, and types of mammogram projections. This approach serves as a starting point for model development of mammography classification tasks. Practitioners can benefit from this work by using the divide and conquer results to select the most suitable deep learning techniques for their case out-of-the-box, thus reducing the need for extensive exploratory experimentation. Multiple techniques are found to provide accuracy gains relative to a general baseline (VGG19 model using uncropped 512 × 512 pixels input images with a dropout rate of 0.2 and a learning rate of 1 × 10(−3)) on the Curated Breast Imaging Subset of DDSM (CBIS-DDSM) dataset. These techniques involve transfer learning pre-trained ImagetNet weights to a MobileNetV2 architecture, with pre-trained weights from a binarised version of the mini Mammography Image Analysis Society (mini-MIAS) dataset applied to the fully connected layers of the model, coupled with using weights to alleviate class imbalance, and splitting CBIS-DDSM samples between images of masses and calcifications. Using these techniques, a 5.6% gain in accuracy over the baseline model was accomplished. Other deep learning techniques from the divide and conquer approach, such as larger image sizes, do not yield increased accuracies without the use of image pre-processing techniques such as Gaussian filtering, histogram equalisation and input cropping.
format Online
Article
Text
id pubmed-10218746
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-102187462023-05-27 A divide and conquer approach to maximise deep learning mammography classification accuracies Jaamour, Adam Myles, Craig Patel, Ashay Chen, Shuen-Jen McMillan, Lewis Harris-Birtill, David PLoS One Research Article Breast cancer claims 11,400 lives on average every year in the UK, making it one of the deadliest diseases. Mammography is the gold standard for detecting early signs of breast cancer, which can help cure the disease during its early stages. However, incorrect mammography diagnoses are common and may harm patients through unnecessary treatments and operations (or a lack of treatment). Therefore, systems that can learn to detect breast cancer on their own could help reduce the number of incorrect interpretations and missed cases. Various deep learning techniques, which can be used to implement a system that learns how to detect instances of breast cancer in mammograms, are explored throughout this paper. Convolution Neural Networks (CNNs) are used as part of a pipeline based on deep learning techniques. A divide and conquer approach is followed to analyse the effects on performance and efficiency when utilising diverse deep learning techniques such as varying network architectures (VGG19, ResNet50, InceptionV3, DenseNet121, MobileNetV2), class weights, input sizes, image ratios, pre-processing techniques, transfer learning, dropout rates, and types of mammogram projections. This approach serves as a starting point for model development of mammography classification tasks. Practitioners can benefit from this work by using the divide and conquer results to select the most suitable deep learning techniques for their case out-of-the-box, thus reducing the need for extensive exploratory experimentation. Multiple techniques are found to provide accuracy gains relative to a general baseline (VGG19 model using uncropped 512 × 512 pixels input images with a dropout rate of 0.2 and a learning rate of 1 × 10(−3)) on the Curated Breast Imaging Subset of DDSM (CBIS-DDSM) dataset. These techniques involve transfer learning pre-trained ImagetNet weights to a MobileNetV2 architecture, with pre-trained weights from a binarised version of the mini Mammography Image Analysis Society (mini-MIAS) dataset applied to the fully connected layers of the model, coupled with using weights to alleviate class imbalance, and splitting CBIS-DDSM samples between images of masses and calcifications. Using these techniques, a 5.6% gain in accuracy over the baseline model was accomplished. Other deep learning techniques from the divide and conquer approach, such as larger image sizes, do not yield increased accuracies without the use of image pre-processing techniques such as Gaussian filtering, histogram equalisation and input cropping. Public Library of Science 2023-05-26 /pmc/articles/PMC10218746/ /pubmed/37235566 http://dx.doi.org/10.1371/journal.pone.0280841 Text en © 2023 Jaamour et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Jaamour, Adam
Myles, Craig
Patel, Ashay
Chen, Shuen-Jen
McMillan, Lewis
Harris-Birtill, David
A divide and conquer approach to maximise deep learning mammography classification accuracies
title A divide and conquer approach to maximise deep learning mammography classification accuracies
title_full A divide and conquer approach to maximise deep learning mammography classification accuracies
title_fullStr A divide and conquer approach to maximise deep learning mammography classification accuracies
title_full_unstemmed A divide and conquer approach to maximise deep learning mammography classification accuracies
title_short A divide and conquer approach to maximise deep learning mammography classification accuracies
title_sort divide and conquer approach to maximise deep learning mammography classification accuracies
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10218746/
https://www.ncbi.nlm.nih.gov/pubmed/37235566
http://dx.doi.org/10.1371/journal.pone.0280841
work_keys_str_mv AT jaamouradam adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT mylescraig adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT patelashay adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT chenshuenjen adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT mcmillanlewis adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT harrisbirtilldavid adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT jaamouradam divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT mylescraig divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT patelashay divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT chenshuenjen divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT mcmillanlewis divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies
AT harrisbirtilldavid divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies