Cargando…

Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT

Cervical cell classification has important clinical significance in cervical cancer screening at early stages. However, there are fewer public cervical cancer smear cell datasets, the weights of each classes’ samples are unbalanced, the image quality is uneven, and the classification research result...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Chen, Shuai, Renjun, Ma, Li, Liu, Wenjia, Wu, Menglin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8933771/
https://www.ncbi.nlm.nih.gov/pubmed/35342326
http://dx.doi.org/10.1007/s11042-022-12670-0
_version_ 1784671729335926784
author Zhao, Chen
Shuai, Renjun
Ma, Li
Liu, Wenjia
Wu, Menglin
author_facet Zhao, Chen
Shuai, Renjun
Ma, Li
Liu, Wenjia
Wu, Menglin
author_sort Zhao, Chen
collection PubMed
description Cervical cell classification has important clinical significance in cervical cancer screening at early stages. However, there are fewer public cervical cancer smear cell datasets, the weights of each classes’ samples are unbalanced, the image quality is uneven, and the classification research results based on CNN tend to overfit. To solve the above problems, we propose a cervical cell image generation model based on taming transformers (CCG-taming transformers) to provide high-quality cervical cancer datasets with sufficient samples and balanced weights, we improve the encoder structure by introducing SE-block and MultiRes-block to improve the ability to extract information from cervical cancer cells images; we introduce Layer Normlization to standardize the data, which is convenient for the subsequent non-linear processing of the data by the ReLU activation function in feed forward; we also introduce SMOTE-Tomek Links to balance the source data set and the number of samples and weights of the images we use Tokens-to-Token Vision Transformers (T2T-ViT) combing transfer learning to classify the cervical cancer smear cell image dataset to improve the classification performance. Classification experiments using the model proposed in this paper are performed on three public cervical cancer datasets, the classification accuracy in the liquid-based cytology Pap smear dataset (4-class), SIPAKMeD (5-class), and Herlev (7-class) are 98.79%, 99.58%, and 99.88%, respectively. The quality of the images we generated on these three data sets is very close to the source data set, the final averaged inception score (IS), Fréchet inception distance (FID), Recall and Precision are 3.75, 0.71, 0.32 and 0.65 respectively. Our method improves the accuracy of cervical cancer smear cell classification, provides more cervical cell sample images for cervical cancer-related research, and assists gynecologists to judge and diagnose different types of cervical cancer cells and analyze cervical cancer cells at different stages, which are difficult to distinguish. This paper applies the transformer to the generation and recognition of cervical cancer cell images for the first time. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s11042-022-12670-0.
format Online
Article
Text
id pubmed-8933771
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-89337712022-03-21 Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT Zhao, Chen Shuai, Renjun Ma, Li Liu, Wenjia Wu, Menglin Multimed Tools Appl Article Cervical cell classification has important clinical significance in cervical cancer screening at early stages. However, there are fewer public cervical cancer smear cell datasets, the weights of each classes’ samples are unbalanced, the image quality is uneven, and the classification research results based on CNN tend to overfit. To solve the above problems, we propose a cervical cell image generation model based on taming transformers (CCG-taming transformers) to provide high-quality cervical cancer datasets with sufficient samples and balanced weights, we improve the encoder structure by introducing SE-block and MultiRes-block to improve the ability to extract information from cervical cancer cells images; we introduce Layer Normlization to standardize the data, which is convenient for the subsequent non-linear processing of the data by the ReLU activation function in feed forward; we also introduce SMOTE-Tomek Links to balance the source data set and the number of samples and weights of the images we use Tokens-to-Token Vision Transformers (T2T-ViT) combing transfer learning to classify the cervical cancer smear cell image dataset to improve the classification performance. Classification experiments using the model proposed in this paper are performed on three public cervical cancer datasets, the classification accuracy in the liquid-based cytology Pap smear dataset (4-class), SIPAKMeD (5-class), and Herlev (7-class) are 98.79%, 99.58%, and 99.88%, respectively. The quality of the images we generated on these three data sets is very close to the source data set, the final averaged inception score (IS), Fréchet inception distance (FID), Recall and Precision are 3.75, 0.71, 0.32 and 0.65 respectively. Our method improves the accuracy of cervical cancer smear cell classification, provides more cervical cell sample images for cervical cancer-related research, and assists gynecologists to judge and diagnose different types of cervical cancer cells and analyze cervical cancer cells at different stages, which are difficult to distinguish. This paper applies the transformer to the generation and recognition of cervical cancer cell images for the first time. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s11042-022-12670-0. Springer US 2022-03-19 2022 /pmc/articles/PMC8933771/ /pubmed/35342326 http://dx.doi.org/10.1007/s11042-022-12670-0 Text en © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Zhao, Chen
Shuai, Renjun
Ma, Li
Liu, Wenjia
Wu, Menglin
Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT
title Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT
title_full Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT
title_fullStr Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT
title_full_unstemmed Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT
title_short Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT
title_sort improving cervical cancer classification with imbalanced datasets combining taming transformers with t2t-vit
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8933771/
https://www.ncbi.nlm.nih.gov/pubmed/35342326
http://dx.doi.org/10.1007/s11042-022-12670-0
work_keys_str_mv AT zhaochen improvingcervicalcancerclassificationwithimbalanceddatasetscombiningtamingtransformerswitht2tvit
AT shuairenjun improvingcervicalcancerclassificationwithimbalanceddatasetscombiningtamingtransformerswitht2tvit
AT mali improvingcervicalcancerclassificationwithimbalanceddatasetscombiningtamingtransformerswitht2tvit
AT liuwenjia improvingcervicalcancerclassificationwithimbalanceddatasetscombiningtamingtransformerswitht2tvit
AT wumenglin improvingcervicalcancerclassificationwithimbalanceddatasetscombiningtamingtransformerswitht2tvit