Cargando…

X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data

This paper addresses the problem of semi-supervised transfer learning with limited cross-modality data in remote sensing. A large amount of multi-modal earth observation images, such as multispectral imagery (MSI) or synthetic aperture radar (SAR) data, are openly available on a global scale, enabli...

Descripción completa

Detalles Bibliográficos
Autores principales: Hong, Danfeng, Yokoya, Naoto, Xia, Gui-Song, Chanussot, Jocelyn, Zhu, Xiao Xiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7453915/
https://www.ncbi.nlm.nih.gov/pubmed/32904376
http://dx.doi.org/10.1016/j.isprsjprs.2020.06.014
_version_ 1783575436526092288
author Hong, Danfeng
Yokoya, Naoto
Xia, Gui-Song
Chanussot, Jocelyn
Zhu, Xiao Xiang
author_facet Hong, Danfeng
Yokoya, Naoto
Xia, Gui-Song
Chanussot, Jocelyn
Zhu, Xiao Xiang
author_sort Hong, Danfeng
collection PubMed
description This paper addresses the problem of semi-supervised transfer learning with limited cross-modality data in remote sensing. A large amount of multi-modal earth observation images, such as multispectral imagery (MSI) or synthetic aperture radar (SAR) data, are openly available on a global scale, enabling parsing global urban scenes through remote sensing imagery. However, their ability in identifying materials (pixel-wise classification) remains limited, due to the noisy collection environment and poor discriminative information as well as limited number of well-annotated training images. To this end, we propose a novel cross-modal deep-learning framework, called X-ModalNet, with three well-designed modules: self-adversarial module, interactive learning module, and label propagation module, by learning to transfer more discriminative information from a small-scale hyperspectral image (HSI) into the classification task using a large-scale MSI or SAR data. Significantly, X-ModalNet generalizes well, owing to propagating labels on an updatable graph constructed by high-level features on the top of the network, yielding semi-supervised cross-modality learning. We evaluate X-ModalNet on two multi-modal remote sensing datasets (HSI-MSI and HSI-SAR) and achieve a significant improvement in comparison with several state-of-the-art methods.
format Online
Article
Text
id pubmed-7453915
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-74539152020-09-02 X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data Hong, Danfeng Yokoya, Naoto Xia, Gui-Song Chanussot, Jocelyn Zhu, Xiao Xiang ISPRS J Photogramm Remote Sens Article This paper addresses the problem of semi-supervised transfer learning with limited cross-modality data in remote sensing. A large amount of multi-modal earth observation images, such as multispectral imagery (MSI) or synthetic aperture radar (SAR) data, are openly available on a global scale, enabling parsing global urban scenes through remote sensing imagery. However, their ability in identifying materials (pixel-wise classification) remains limited, due to the noisy collection environment and poor discriminative information as well as limited number of well-annotated training images. To this end, we propose a novel cross-modal deep-learning framework, called X-ModalNet, with three well-designed modules: self-adversarial module, interactive learning module, and label propagation module, by learning to transfer more discriminative information from a small-scale hyperspectral image (HSI) into the classification task using a large-scale MSI or SAR data. Significantly, X-ModalNet generalizes well, owing to propagating labels on an updatable graph constructed by high-level features on the top of the network, yielding semi-supervised cross-modality learning. We evaluate X-ModalNet on two multi-modal remote sensing datasets (HSI-MSI and HSI-SAR) and achieve a significant improvement in comparison with several state-of-the-art methods. Elsevier 2020-09 /pmc/articles/PMC7453915/ /pubmed/32904376 http://dx.doi.org/10.1016/j.isprsjprs.2020.06.014 Text en © 2020 The Authors. Published by Elsevier B.V. on behalf of International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS). http://creativecommons.org/licenses/by/4.0/ This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Hong, Danfeng
Yokoya, Naoto
Xia, Gui-Song
Chanussot, Jocelyn
Zhu, Xiao Xiang
X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
title X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
title_full X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
title_fullStr X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
title_full_unstemmed X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
title_short X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data
title_sort x-modalnet: a semi-supervised deep cross-modal network for classification of remote sensing data
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7453915/
https://www.ncbi.nlm.nih.gov/pubmed/32904376
http://dx.doi.org/10.1016/j.isprsjprs.2020.06.014
work_keys_str_mv AT hongdanfeng xmodalnetasemisuperviseddeepcrossmodalnetworkforclassificationofremotesensingdata
AT yokoyanaoto xmodalnetasemisuperviseddeepcrossmodalnetworkforclassificationofremotesensingdata
AT xiaguisong xmodalnetasemisuperviseddeepcrossmodalnetworkforclassificationofremotesensingdata
AT chanussotjocelyn xmodalnetasemisuperviseddeepcrossmodalnetworkforclassificationofremotesensingdata
AT zhuxiaoxiang xmodalnetasemisuperviseddeepcrossmodalnetworkforclassificationofremotesensingdata