Cargando…

Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer

Medical image segmentation is a vital healthcare endeavor requiring precise and efficient models for appropriate diagnosis and treatment. Vision transformer (ViT)-based segmentation models have shown great performance in accomplishing this task. However, to build a powerful backbone, the self-attent...

Descripción completa

Detalles Bibliográficos
Autores principales: Saeed, Numan, Ridzuan, Muhammad, Majzoub, Roba Al, Yaqub, Mohammad
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10376048/
https://www.ncbi.nlm.nih.gov/pubmed/37508906
http://dx.doi.org/10.3390/bioengineering10070879
_version_ 1785079175025000448
author Saeed, Numan
Ridzuan, Muhammad
Majzoub, Roba Al
Yaqub, Mohammad
author_facet Saeed, Numan
Ridzuan, Muhammad
Majzoub, Roba Al
Yaqub, Mohammad
author_sort Saeed, Numan
collection PubMed
description Medical image segmentation is a vital healthcare endeavor requiring precise and efficient models for appropriate diagnosis and treatment. Vision transformer (ViT)-based segmentation models have shown great performance in accomplishing this task. However, to build a powerful backbone, the self-attention block of ViT requires large-scale pre-training data. The present method of modifying pre-trained models entails updating all or some of the backbone parameters. This paper proposes a novel fine-tuning strategy for adapting a pretrained transformer-based segmentation model on data from a new medical center. This method introduces a small number of learnable parameters, termed prompts, into the input space (less than 1% of model parameters) while keeping the rest of the model parameters frozen. Extensive studies employing data from new unseen medical centers show that the prompt-based fine-tuning of medical segmentation models provides excellent performance regarding the new-center data with a negligible drop regarding the old centers. Additionally, our strategy delivers great accuracy with minimum re-training on new-center data, significantly decreasing the computational and time costs of fine-tuning pre-trained models. Our source code will be made publicly available.
format Online
Article
Text
id pubmed-10376048
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-103760482023-07-29 Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer Saeed, Numan Ridzuan, Muhammad Majzoub, Roba Al Yaqub, Mohammad Bioengineering (Basel) Article Medical image segmentation is a vital healthcare endeavor requiring precise and efficient models for appropriate diagnosis and treatment. Vision transformer (ViT)-based segmentation models have shown great performance in accomplishing this task. However, to build a powerful backbone, the self-attention block of ViT requires large-scale pre-training data. The present method of modifying pre-trained models entails updating all or some of the backbone parameters. This paper proposes a novel fine-tuning strategy for adapting a pretrained transformer-based segmentation model on data from a new medical center. This method introduces a small number of learnable parameters, termed prompts, into the input space (less than 1% of model parameters) while keeping the rest of the model parameters frozen. Extensive studies employing data from new unseen medical centers show that the prompt-based fine-tuning of medical segmentation models provides excellent performance regarding the new-center data with a negligible drop regarding the old centers. Additionally, our strategy delivers great accuracy with minimum re-training on new-center data, significantly decreasing the computational and time costs of fine-tuning pre-trained models. Our source code will be made publicly available. MDPI 2023-07-24 /pmc/articles/PMC10376048/ /pubmed/37508906 http://dx.doi.org/10.3390/bioengineering10070879 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Saeed, Numan
Ridzuan, Muhammad
Majzoub, Roba Al
Yaqub, Mohammad
Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
title Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
title_full Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
title_fullStr Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
title_full_unstemmed Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
title_short Prompt-Based Tuning of Transformer Models for Multi-Center Medical Image Segmentation of Head and Neck Cancer
title_sort prompt-based tuning of transformer models for multi-center medical image segmentation of head and neck cancer
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10376048/
https://www.ncbi.nlm.nih.gov/pubmed/37508906
http://dx.doi.org/10.3390/bioengineering10070879
work_keys_str_mv AT saeednuman promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer
AT ridzuanmuhammad promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer
AT majzoubrobaal promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer
AT yaqubmohammad promptbasedtuningoftransformermodelsformulticentermedicalimagesegmentationofheadandneckcancer