Cargando…

Center transfer for supervised domain adaptation

Domain adaptation (DA) is a popular strategy for pattern recognition and classification tasks. It leverages a large amount of data from the source domain to help train the model applied in the target domain. Supervised domain adaptation (SDA) approaches are desirable when only few labeled samples fr...

Descripción completa

Detalles Bibliográficos
Autores principales: Huang, Xiuyu, Zhou, Nan, Huang, Jian, Zhang, Huaidong, Pedrycz, Witold, Choi, Kup-Sze
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9878501/
https://www.ncbi.nlm.nih.gov/pubmed/36718382
http://dx.doi.org/10.1007/s10489-022-04414-2
_version_ 1784878497694482432
author Huang, Xiuyu
Zhou, Nan
Huang, Jian
Zhang, Huaidong
Pedrycz, Witold
Choi, Kup-Sze
author_facet Huang, Xiuyu
Zhou, Nan
Huang, Jian
Zhang, Huaidong
Pedrycz, Witold
Choi, Kup-Sze
author_sort Huang, Xiuyu
collection PubMed
description Domain adaptation (DA) is a popular strategy for pattern recognition and classification tasks. It leverages a large amount of data from the source domain to help train the model applied in the target domain. Supervised domain adaptation (SDA) approaches are desirable when only few labeled samples from the target domain are available. They can be easily adopted in many real-world applications where data collection is expensive. In this study, we propose a new supervision signal, namely center transfer loss (CTL), to efficiently align features under the SDA setting in the deep learning (DL) field. Unlike most previous SDA methods that rely on pairing up training samples, the proposed loss is trainable only using one-stream input based on the mini-batch strategy. The CTL exhibits two main functionalities in training to increase the performance of DL models, i.e., domain alignment and increasing the feature’s discriminative power. The hyper-parameter to balance these two functionalities is waived in CTL, which is the second improvement from the previous approaches. Extensive experiments completed on well-known public datasets show that the proposed method performs better than recent state-of-the-art approaches.
format Online
Article
Text
id pubmed-9878501
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-98785012023-01-26 Center transfer for supervised domain adaptation Huang, Xiuyu Zhou, Nan Huang, Jian Zhang, Huaidong Pedrycz, Witold Choi, Kup-Sze Appl Intell (Dordr) Article Domain adaptation (DA) is a popular strategy for pattern recognition and classification tasks. It leverages a large amount of data from the source domain to help train the model applied in the target domain. Supervised domain adaptation (SDA) approaches are desirable when only few labeled samples from the target domain are available. They can be easily adopted in many real-world applications where data collection is expensive. In this study, we propose a new supervision signal, namely center transfer loss (CTL), to efficiently align features under the SDA setting in the deep learning (DL) field. Unlike most previous SDA methods that rely on pairing up training samples, the proposed loss is trainable only using one-stream input based on the mini-batch strategy. The CTL exhibits two main functionalities in training to increase the performance of DL models, i.e., domain alignment and increasing the feature’s discriminative power. The hyper-parameter to balance these two functionalities is waived in CTL, which is the second improvement from the previous approaches. Extensive experiments completed on well-known public datasets show that the proposed method performs better than recent state-of-the-art approaches. Springer US 2023-01-26 /pmc/articles/PMC9878501/ /pubmed/36718382 http://dx.doi.org/10.1007/s10489-022-04414-2 Text en © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023, Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Huang, Xiuyu
Zhou, Nan
Huang, Jian
Zhang, Huaidong
Pedrycz, Witold
Choi, Kup-Sze
Center transfer for supervised domain adaptation
title Center transfer for supervised domain adaptation
title_full Center transfer for supervised domain adaptation
title_fullStr Center transfer for supervised domain adaptation
title_full_unstemmed Center transfer for supervised domain adaptation
title_short Center transfer for supervised domain adaptation
title_sort center transfer for supervised domain adaptation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9878501/
https://www.ncbi.nlm.nih.gov/pubmed/36718382
http://dx.doi.org/10.1007/s10489-022-04414-2
work_keys_str_mv AT huangxiuyu centertransferforsuperviseddomainadaptation
AT zhounan centertransferforsuperviseddomainadaptation
AT huangjian centertransferforsuperviseddomainadaptation
AT zhanghuaidong centertransferforsuperviseddomainadaptation
AT pedryczwitold centertransferforsuperviseddomainadaptation
AT choikupsze centertransferforsuperviseddomainadaptation