Cargando…

Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images

The United Nations predicts that by 2050, the world’s total population will increase to 9.15 billion, but the per capita cropland will drop to 0.151°hm(2). The acceleration of urbanization often comes at the expense of the encroachment of cropland, the unplanned expansion of urban area has adversely...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Junshu, Cai, Mingrui, Gu, Yifan, Liu, Zhen, Li, Xiaoxin, Han, Yuxing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9486080/
https://www.ncbi.nlm.nih.gov/pubmed/36147239
http://dx.doi.org/10.3389/fpls.2022.993961
_version_ 1784792198521290752
author Wang, Junshu
Cai, Mingrui
Gu, Yifan
Liu, Zhen
Li, Xiaoxin
Han, Yuxing
author_facet Wang, Junshu
Cai, Mingrui
Gu, Yifan
Liu, Zhen
Li, Xiaoxin
Han, Yuxing
author_sort Wang, Junshu
collection PubMed
description The United Nations predicts that by 2050, the world’s total population will increase to 9.15 billion, but the per capita cropland will drop to 0.151°hm(2). The acceleration of urbanization often comes at the expense of the encroachment of cropland, the unplanned expansion of urban area has adversely affected cultivation. Therefore, the automatic extraction of buildings, which are the main carriers of urban population activities, in remote sensing images has become a more meaningful cropland observation task. To solve the shortcomings of traditional building extraction methods such as insufficient utilization of image information, relying on manual characterization, etc. A U-Net based deep learning building extraction model is proposed and named AttsegGAN. This study proposes an adversarial loss based on the Generative Adversarial Network in terms of training strategy, and the additionally trained learnable discriminator is used as a distance measurer for the two probability distributions of ground truth P(data) and prediction P(g). In addition, for the sharpness of the building edge, the Sobel edge loss based on the Sobel operator is weighted and jointly participated in the training. In WHU building dataset, this study applies the components and strategies step by step, and verifies their effectiveness. Furthermore, the addition of the attention module is also subjected to ablation experiments and the final framework is determined. Compared with the original, AttsegGAN improved by 0.0062, 0.0027, and 0.0055 on Acc, F1, and IoU respectively after adopting all improvements. In the comparative experiment. AttsegGAN is compared with state-of-the-arts including U-Net, DeeplabV3+, PSPNet, and DANet on both WHU and Massachusetts building dataset. In WHU dataset, AttsegGAN achieved 0.9875, 0.9435, and 0.8907 on Acc, F1, and IoU, surpassed U-Net by 0.0260, 0.1183, and 0.1883, respectively, demonstrated the effectiveness of the proposed components in a similar hourglass structure. In Massachusetts dataset, AttsegGAN also surpassed state-of-the-arts, achieved 0.9395, 0.8328, and 0.7130 on Acc, F1, and IoU, respectively, it improved IoU by 0.0412 over the second-ranked PSPNet, and it was 0.0025 and 0.0101 higher than the second place in Acc and F1.
format Online
Article
Text
id pubmed-9486080
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-94860802022-09-21 Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images Wang, Junshu Cai, Mingrui Gu, Yifan Liu, Zhen Li, Xiaoxin Han, Yuxing Front Plant Sci Plant Science The United Nations predicts that by 2050, the world’s total population will increase to 9.15 billion, but the per capita cropland will drop to 0.151°hm(2). The acceleration of urbanization often comes at the expense of the encroachment of cropland, the unplanned expansion of urban area has adversely affected cultivation. Therefore, the automatic extraction of buildings, which are the main carriers of urban population activities, in remote sensing images has become a more meaningful cropland observation task. To solve the shortcomings of traditional building extraction methods such as insufficient utilization of image information, relying on manual characterization, etc. A U-Net based deep learning building extraction model is proposed and named AttsegGAN. This study proposes an adversarial loss based on the Generative Adversarial Network in terms of training strategy, and the additionally trained learnable discriminator is used as a distance measurer for the two probability distributions of ground truth P(data) and prediction P(g). In addition, for the sharpness of the building edge, the Sobel edge loss based on the Sobel operator is weighted and jointly participated in the training. In WHU building dataset, this study applies the components and strategies step by step, and verifies their effectiveness. Furthermore, the addition of the attention module is also subjected to ablation experiments and the final framework is determined. Compared with the original, AttsegGAN improved by 0.0062, 0.0027, and 0.0055 on Acc, F1, and IoU respectively after adopting all improvements. In the comparative experiment. AttsegGAN is compared with state-of-the-arts including U-Net, DeeplabV3+, PSPNet, and DANet on both WHU and Massachusetts building dataset. In WHU dataset, AttsegGAN achieved 0.9875, 0.9435, and 0.8907 on Acc, F1, and IoU, surpassed U-Net by 0.0260, 0.1183, and 0.1883, respectively, demonstrated the effectiveness of the proposed components in a similar hourglass structure. In Massachusetts dataset, AttsegGAN also surpassed state-of-the-arts, achieved 0.9395, 0.8328, and 0.7130 on Acc, F1, and IoU, respectively, it improved IoU by 0.0412 over the second-ranked PSPNet, and it was 0.0025 and 0.0101 higher than the second place in Acc and F1. Frontiers Media S.A. 2022-09-06 /pmc/articles/PMC9486080/ /pubmed/36147239 http://dx.doi.org/10.3389/fpls.2022.993961 Text en Copyright © 2022 Wang, Cai, Gu, Liu, Li and Han. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Plant Science
Wang, Junshu
Cai, Mingrui
Gu, Yifan
Liu, Zhen
Li, Xiaoxin
Han, Yuxing
Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
title Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
title_full Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
title_fullStr Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
title_full_unstemmed Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
title_short Cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
title_sort cropland encroachment detection via dual attention and multi-loss based building extraction in remote sensing images
topic Plant Science
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9486080/
https://www.ncbi.nlm.nih.gov/pubmed/36147239
http://dx.doi.org/10.3389/fpls.2022.993961
work_keys_str_mv AT wangjunshu croplandencroachmentdetectionviadualattentionandmultilossbasedbuildingextractioninremotesensingimages
AT caimingrui croplandencroachmentdetectionviadualattentionandmultilossbasedbuildingextractioninremotesensingimages
AT guyifan croplandencroachmentdetectionviadualattentionandmultilossbasedbuildingextractioninremotesensingimages
AT liuzhen croplandencroachmentdetectionviadualattentionandmultilossbasedbuildingextractioninremotesensingimages
AT lixiaoxin croplandencroachmentdetectionviadualattentionandmultilossbasedbuildingextractioninremotesensingimages
AT hanyuxing croplandencroachmentdetectionviadualattentionandmultilossbasedbuildingextractioninremotesensingimages