Cargando…
A divide and conquer approach to maximise deep learning mammography classification accuracies
Breast cancer claims 11,400 lives on average every year in the UK, making it one of the deadliest diseases. Mammography is the gold standard for detecting early signs of breast cancer, which can help cure the disease during its early stages. However, incorrect mammography diagnoses are common and ma...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10218746/ https://www.ncbi.nlm.nih.gov/pubmed/37235566 http://dx.doi.org/10.1371/journal.pone.0280841 |
_version_ | 1785048846993195008 |
---|---|
author | Jaamour, Adam Myles, Craig Patel, Ashay Chen, Shuen-Jen McMillan, Lewis Harris-Birtill, David |
author_facet | Jaamour, Adam Myles, Craig Patel, Ashay Chen, Shuen-Jen McMillan, Lewis Harris-Birtill, David |
author_sort | Jaamour, Adam |
collection | PubMed |
description | Breast cancer claims 11,400 lives on average every year in the UK, making it one of the deadliest diseases. Mammography is the gold standard for detecting early signs of breast cancer, which can help cure the disease during its early stages. However, incorrect mammography diagnoses are common and may harm patients through unnecessary treatments and operations (or a lack of treatment). Therefore, systems that can learn to detect breast cancer on their own could help reduce the number of incorrect interpretations and missed cases. Various deep learning techniques, which can be used to implement a system that learns how to detect instances of breast cancer in mammograms, are explored throughout this paper. Convolution Neural Networks (CNNs) are used as part of a pipeline based on deep learning techniques. A divide and conquer approach is followed to analyse the effects on performance and efficiency when utilising diverse deep learning techniques such as varying network architectures (VGG19, ResNet50, InceptionV3, DenseNet121, MobileNetV2), class weights, input sizes, image ratios, pre-processing techniques, transfer learning, dropout rates, and types of mammogram projections. This approach serves as a starting point for model development of mammography classification tasks. Practitioners can benefit from this work by using the divide and conquer results to select the most suitable deep learning techniques for their case out-of-the-box, thus reducing the need for extensive exploratory experimentation. Multiple techniques are found to provide accuracy gains relative to a general baseline (VGG19 model using uncropped 512 × 512 pixels input images with a dropout rate of 0.2 and a learning rate of 1 × 10(−3)) on the Curated Breast Imaging Subset of DDSM (CBIS-DDSM) dataset. These techniques involve transfer learning pre-trained ImagetNet weights to a MobileNetV2 architecture, with pre-trained weights from a binarised version of the mini Mammography Image Analysis Society (mini-MIAS) dataset applied to the fully connected layers of the model, coupled with using weights to alleviate class imbalance, and splitting CBIS-DDSM samples between images of masses and calcifications. Using these techniques, a 5.6% gain in accuracy over the baseline model was accomplished. Other deep learning techniques from the divide and conquer approach, such as larger image sizes, do not yield increased accuracies without the use of image pre-processing techniques such as Gaussian filtering, histogram equalisation and input cropping. |
format | Online Article Text |
id | pubmed-10218746 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-102187462023-05-27 A divide and conquer approach to maximise deep learning mammography classification accuracies Jaamour, Adam Myles, Craig Patel, Ashay Chen, Shuen-Jen McMillan, Lewis Harris-Birtill, David PLoS One Research Article Breast cancer claims 11,400 lives on average every year in the UK, making it one of the deadliest diseases. Mammography is the gold standard for detecting early signs of breast cancer, which can help cure the disease during its early stages. However, incorrect mammography diagnoses are common and may harm patients through unnecessary treatments and operations (or a lack of treatment). Therefore, systems that can learn to detect breast cancer on their own could help reduce the number of incorrect interpretations and missed cases. Various deep learning techniques, which can be used to implement a system that learns how to detect instances of breast cancer in mammograms, are explored throughout this paper. Convolution Neural Networks (CNNs) are used as part of a pipeline based on deep learning techniques. A divide and conquer approach is followed to analyse the effects on performance and efficiency when utilising diverse deep learning techniques such as varying network architectures (VGG19, ResNet50, InceptionV3, DenseNet121, MobileNetV2), class weights, input sizes, image ratios, pre-processing techniques, transfer learning, dropout rates, and types of mammogram projections. This approach serves as a starting point for model development of mammography classification tasks. Practitioners can benefit from this work by using the divide and conquer results to select the most suitable deep learning techniques for their case out-of-the-box, thus reducing the need for extensive exploratory experimentation. Multiple techniques are found to provide accuracy gains relative to a general baseline (VGG19 model using uncropped 512 × 512 pixels input images with a dropout rate of 0.2 and a learning rate of 1 × 10(−3)) on the Curated Breast Imaging Subset of DDSM (CBIS-DDSM) dataset. These techniques involve transfer learning pre-trained ImagetNet weights to a MobileNetV2 architecture, with pre-trained weights from a binarised version of the mini Mammography Image Analysis Society (mini-MIAS) dataset applied to the fully connected layers of the model, coupled with using weights to alleviate class imbalance, and splitting CBIS-DDSM samples between images of masses and calcifications. Using these techniques, a 5.6% gain in accuracy over the baseline model was accomplished. Other deep learning techniques from the divide and conquer approach, such as larger image sizes, do not yield increased accuracies without the use of image pre-processing techniques such as Gaussian filtering, histogram equalisation and input cropping. Public Library of Science 2023-05-26 /pmc/articles/PMC10218746/ /pubmed/37235566 http://dx.doi.org/10.1371/journal.pone.0280841 Text en © 2023 Jaamour et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Jaamour, Adam Myles, Craig Patel, Ashay Chen, Shuen-Jen McMillan, Lewis Harris-Birtill, David A divide and conquer approach to maximise deep learning mammography classification accuracies |
title | A divide and conquer approach to maximise deep learning mammography classification accuracies |
title_full | A divide and conquer approach to maximise deep learning mammography classification accuracies |
title_fullStr | A divide and conquer approach to maximise deep learning mammography classification accuracies |
title_full_unstemmed | A divide and conquer approach to maximise deep learning mammography classification accuracies |
title_short | A divide and conquer approach to maximise deep learning mammography classification accuracies |
title_sort | divide and conquer approach to maximise deep learning mammography classification accuracies |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10218746/ https://www.ncbi.nlm.nih.gov/pubmed/37235566 http://dx.doi.org/10.1371/journal.pone.0280841 |
work_keys_str_mv | AT jaamouradam adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT mylescraig adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT patelashay adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT chenshuenjen adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT mcmillanlewis adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT harrisbirtilldavid adivideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT jaamouradam divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT mylescraig divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT patelashay divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT chenshuenjen divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT mcmillanlewis divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies AT harrisbirtilldavid divideandconquerapproachtomaximisedeeplearningmammographyclassificationaccuracies |