Cargando…

Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features

Land use classification is a fundamental task of information extraction from remote sensing imagery. Semantic segmentation based on deep convolutional neural networks (DCNNs) has shown outstanding performance in this task. However, these methods are still affected by the loss of spatial features. In...

Descripción completa

Detalles Bibliográficos
Autores principales: Yao, Xuedong, Yang, Hui, Wu, Yanlan, Wu, Penghai, Wang, Biao, Zhou, Xinxin, Wang, Shuai
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6631649/
https://www.ncbi.nlm.nih.gov/pubmed/31234384
http://dx.doi.org/10.3390/s19122792
_version_ 1783435567258664960
author Yao, Xuedong
Yang, Hui
Wu, Yanlan
Wu, Penghai
Wang, Biao
Zhou, Xinxin
Wang, Shuai
author_facet Yao, Xuedong
Yang, Hui
Wu, Yanlan
Wu, Penghai
Wang, Biao
Zhou, Xinxin
Wang, Shuai
author_sort Yao, Xuedong
collection PubMed
description Land use classification is a fundamental task of information extraction from remote sensing imagery. Semantic segmentation based on deep convolutional neural networks (DCNNs) has shown outstanding performance in this task. However, these methods are still affected by the loss of spatial features. In this study, we proposed a new network, called the dense-coordconv network (DCCN), to reduce the loss of spatial features and strengthen object boundaries. In this network, the coordconv module is introduced into the improved DenseNet architecture to improve spatial information by putting coordinate information into feature maps. The proposed DCCN achieved an obvious performance in terms of the public ISPRS (International Society for Photogrammetry and Remote Sensing) 2D semantic labeling benchmark dataset. Compared with the results of other deep convolutional neural networks (U-net, SegNet, Deeplab-V3), the results of the DCCN method improved a lot and the OA (overall accuracy) and mean F1 score reached 89.48% and 86.89%, respectively. This indicates that the DCCN method can effectively reduce the loss of spatial features and improve the accuracy of semantic segmentation in high resolution remote sensing imagery.
format Online
Article
Text
id pubmed-6631649
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-66316492019-08-19 Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features Yao, Xuedong Yang, Hui Wu, Yanlan Wu, Penghai Wang, Biao Zhou, Xinxin Wang, Shuai Sensors (Basel) Article Land use classification is a fundamental task of information extraction from remote sensing imagery. Semantic segmentation based on deep convolutional neural networks (DCNNs) has shown outstanding performance in this task. However, these methods are still affected by the loss of spatial features. In this study, we proposed a new network, called the dense-coordconv network (DCCN), to reduce the loss of spatial features and strengthen object boundaries. In this network, the coordconv module is introduced into the improved DenseNet architecture to improve spatial information by putting coordinate information into feature maps. The proposed DCCN achieved an obvious performance in terms of the public ISPRS (International Society for Photogrammetry and Remote Sensing) 2D semantic labeling benchmark dataset. Compared with the results of other deep convolutional neural networks (U-net, SegNet, Deeplab-V3), the results of the DCCN method improved a lot and the OA (overall accuracy) and mean F1 score reached 89.48% and 86.89%, respectively. This indicates that the DCCN method can effectively reduce the loss of spatial features and improve the accuracy of semantic segmentation in high resolution remote sensing imagery. MDPI 2019-06-21 /pmc/articles/PMC6631649/ /pubmed/31234384 http://dx.doi.org/10.3390/s19122792 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Yao, Xuedong
Yang, Hui
Wu, Yanlan
Wu, Penghai
Wang, Biao
Zhou, Xinxin
Wang, Shuai
Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features
title Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features
title_full Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features
title_fullStr Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features
title_full_unstemmed Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features
title_short Land Use Classification of the Deep Convolutional Neural Network Method Reducing the Loss of Spatial Features
title_sort land use classification of the deep convolutional neural network method reducing the loss of spatial features
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6631649/
https://www.ncbi.nlm.nih.gov/pubmed/31234384
http://dx.doi.org/10.3390/s19122792
work_keys_str_mv AT yaoxuedong landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures
AT yanghui landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures
AT wuyanlan landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures
AT wupenghai landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures
AT wangbiao landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures
AT zhouxinxin landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures
AT wangshuai landuseclassificationofthedeepconvolutionalneuralnetworkmethodreducingthelossofspatialfeatures