Cargando…
Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
Medical image segmentation is a vital healthcare endeavor requiring precise and efficient models for appropriate diagnosis and treatment. Vision transformer (ViT)-based segmentation models have shown great performance in accomplishing this task. However, to build a powerful backbone, the self-attent...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10376048/ https://www.ncbi.nlm.nih.gov/pubmed/37508906 http://dx.doi.org/10.3390/bioengineering10070879 |
_version_ | 1785079175025000448 |
---|---|
author | Saeed, Numan Ridzuan, Muhammad Majzoub, Roba Al Yaqub, Mohammad |
author_facet | Saeed, Numan Ridzuan, Muhammad Majzoub, Roba Al Yaqub, Mohammad |
author_sort | Saeed, Numan |
collection | PubMed |
description | Medical image segmentation is a vital healthcare endeavor requiring precise and efficient models for appropriate diagnosis and treatment. Vision transformer (ViT)-based segmentation models have shown great performance in accomplishing this task. However, to build a powerful backbone, the self-attention block of ViT requires large-scale pre-training data. The present method of modifying pre-trained models entails updating all or some of the backbone parameters. This paper proposes a novel fine-tuning strategy for adapting a pretrained transformer-based segmentation model on data from a new medical center. This method introduces a small number of learnable parameters, termed prompts, into the input space (less than 1% of model parameters) while keeping the rest of the model parameters frozen. Extensive studies employing data from new unseen medical centers show that the prompt-based fine-tuning of medical segmentation models provides excellent performance regarding the new-center data with a negligible drop regarding the old centers. Additionally, our strategy delivers great accuracy with minimum re-training on new-center data, significantly decreasing the computational and time costs of fine-tuning pre-trained models. Our source code will be made publicly available. |
format | Online Article Text |
id | pubmed-10376048 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-103760482023-07-29 Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer Saeed, Numan Ridzuan, Muhammad Majzoub, Roba Al Yaqub, Mohammad Bioengineering (Basel) Article Medical image segmentation is a vital healthcare endeavor requiring precise and efficient models for appropriate diagnosis and treatment. Vision transformer (ViT)-based segmentation models have shown great performance in accomplishing this task. However, to build a powerful backbone, the self-attention block of ViT requires large-scale pre-training data. The present method of modifying pre-trained models entails updating all or some of the backbone parameters. This paper proposes a novel fine-tuning strategy for adapting a pretrained transformer-based segmentation model on data from a new medical center. This method introduces a small number of learnable parameters, termed prompts, into the input space (less than 1% of model parameters) while keeping the rest of the model parameters frozen. Extensive studies employing data from new unseen medical centers show that the prompt-based fine-tuning of medical segmentation models provides excellent performance regarding the new-center data with a negligible drop regarding the old centers. Additionally, our strategy delivers great accuracy with minimum re-training on new-center data, significantly decreasing the computational and time costs of fine-tuning pre-trained models. Our source code will be made publicly available. MDPI 2023-07-24 /pmc/articles/PMC10376048/ /pubmed/37508906 http://dx.doi.org/10.3390/bioengineering10070879 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Saeed, Numan Ridzuan, Muhammad Majzoub, Roba Al Yaqub, Mohammad Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer |
title | Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer |
title_full | Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer |
title_fullStr | Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer |
title_full_unstemmed | Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer |
title_short | Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer |
title_sort | prompt-based tuning of transformer models for multi-center medical image segmentation of head and neck cancer |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10376048/ https://www.ncbi.nlm.nih.gov/pubmed/37508906 http://dx.doi.org/10.3390/bioengineering10070879 |
work_keys_str_mv | AT saeednuman promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer AT ridzuanmuhammad promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer AT majzoubrobaal promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer AT yaqubmohammad promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer |