Cargando…
Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images
Semantic segmentation with deep learning networks has become an important approach to the extraction of objects from very high-resolution remote sensing images. Vision Transformer networks have shown significant improvements in performance compared to traditional convolutional neural networks (CNNs)...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10255903/ https://www.ncbi.nlm.nih.gov/pubmed/37299892 http://dx.doi.org/10.3390/s23115166 |
_version_ | 1785056985415155712 |
---|---|
author | Song, Jia Zhu, A-Xing Zhu, Yunqiang |
author_facet | Song, Jia Zhu, A-Xing Zhu, Yunqiang |
author_sort | Song, Jia |
collection | PubMed |
description | Semantic segmentation with deep learning networks has become an important approach to the extraction of objects from very high-resolution remote sensing images. Vision Transformer networks have shown significant improvements in performance compared to traditional convolutional neural networks (CNNs) in semantic segmentation. Vision Transformer networks have different architectures to CNNs. Image patches, linear embedding, and multi-head self-attention (MHSA) are several of the main hyperparameters. How we should configure them for the extraction of objects in VHR images and how they affect the accuracy of networks are topics that have not been sufficiently investigated. This article explores the role of vision Transformer networks in the extraction of building footprints from very-high-resolution (VHR) images. Transformer-based models with different hyperparameter values were designed and compared, and their impact on accuracy was analyzed. The results show that smaller image patches and higher-dimension embeddings result in better accuracy. In addition, the Transformer-based network is shown to be scalable and can be trained with general-scale graphics processing units (GPUs) with comparable model sizes and training times to convolutional neural networks while achieving higher accuracy. The study provides valuable insights into the potential of vision Transformer networks in object extraction using VHR images. |
format | Online Article Text |
id | pubmed-10255903 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-102559032023-06-10 Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images Song, Jia Zhu, A-Xing Zhu, Yunqiang Sensors (Basel) Article Semantic segmentation with deep learning networks has become an important approach to the extraction of objects from very high-resolution remote sensing images. Vision Transformer networks have shown significant improvements in performance compared to traditional convolutional neural networks (CNNs) in semantic segmentation. Vision Transformer networks have different architectures to CNNs. Image patches, linear embedding, and multi-head self-attention (MHSA) are several of the main hyperparameters. How we should configure them for the extraction of objects in VHR images and how they affect the accuracy of networks are topics that have not been sufficiently investigated. This article explores the role of vision Transformer networks in the extraction of building footprints from very-high-resolution (VHR) images. Transformer-based models with different hyperparameter values were designed and compared, and their impact on accuracy was analyzed. The results show that smaller image patches and higher-dimension embeddings result in better accuracy. In addition, the Transformer-based network is shown to be scalable and can be trained with general-scale graphics processing units (GPUs) with comparable model sizes and training times to convolutional neural networks while achieving higher accuracy. The study provides valuable insights into the potential of vision Transformer networks in object extraction using VHR images. MDPI 2023-05-29 /pmc/articles/PMC10255903/ /pubmed/37299892 http://dx.doi.org/10.3390/s23115166 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Song, Jia Zhu, A-Xing Zhu, Yunqiang Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images |
title | Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images |
title_full | Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images |
title_fullStr | Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images |
title_full_unstemmed | Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images |
title_short | Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images |
title_sort | transformer-based semantic segmentation for extraction of building footprints from very-high-resolution images |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10255903/ https://www.ncbi.nlm.nih.gov/pubmed/37299892 http://dx.doi.org/10.3390/s23115166 |
work_keys_str_mv | AT songjia transformerbasedsemanticsegmentationforextractionofbuildingfootprintsfromveryhighresolutionimages AT zhuaxing transformerbasedsemanticsegmentationforextractionofbuildingfootprintsfromveryhighresolutionimages AT zhuyunqiang transformerbasedsemanticsegmentationforextractionofbuildingfootprintsfromveryhighresolutionimages |