Cargando…
Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning
Semi-supervised learning is a learning pattern that can utilize labeled data and unlabeled data to train deep neural networks. In semi-supervised learning methods, self-training-based methods do not depend on a data augmentation strategy and have better generalization ability. However, their perform...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10145307/ https://www.ncbi.nlm.nih.gov/pubmed/37112285 http://dx.doi.org/10.3390/s23083944 |
_version_ | 1785034302074912768 |
---|---|
author | Wang, Yifan Huang, Yan Wang, Qicong Zhao, Chong Zhang, Zhenchang Chen, Jian |
author_facet | Wang, Yifan Huang, Yan Wang, Qicong Zhao, Chong Zhang, Zhenchang Chen, Jian |
author_sort | Wang, Yifan |
collection | PubMed |
description | Semi-supervised learning is a learning pattern that can utilize labeled data and unlabeled data to train deep neural networks. In semi-supervised learning methods, self-training-based methods do not depend on a data augmentation strategy and have better generalization ability. However, their performance is limited by the accuracy of predicted pseudo-labels. In this paper, we propose to reduce the noise in the pseudo-labels from two aspects: the accuracy of predictions and the confidence of the predictions. For the first aspect, we propose a similarity graph structure learning (SGSL) model that considers the correlation between unlabeled and labeled samples, which facilitates the learning of more discriminative features and, thus, obtains more accurate predictions. For the second aspect, we propose an uncertainty-based graph convolutional network (UGCN), which can aggregate similar features based on the learned graph structure in the training phase, making the features more discriminative. It can also output the uncertainty of predictions in the pseudo-label generation phase, generating pseudo-labels only for unlabeled samples with low uncertainty; thus, reducing the noise in the pseudo-labels. Further, a positive and negative self-training framework is proposed, which combines the proposed SGSL model and UGCN into the self-training framework for end-to-end training. In addition, in order to introduce more supervised signals in the self-training process, negative pseudo-labels are generated for unlabeled samples with low prediction confidence, and then the positive and negative pseudo-labeled samples are trained together with a small number of labeled samples to improve the performance of semi-supervised learning. The code is available upon request. |
format | Online Article Text |
id | pubmed-10145307 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-101453072023-04-29 Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning Wang, Yifan Huang, Yan Wang, Qicong Zhao, Chong Zhang, Zhenchang Chen, Jian Sensors (Basel) Article Semi-supervised learning is a learning pattern that can utilize labeled data and unlabeled data to train deep neural networks. In semi-supervised learning methods, self-training-based methods do not depend on a data augmentation strategy and have better generalization ability. However, their performance is limited by the accuracy of predicted pseudo-labels. In this paper, we propose to reduce the noise in the pseudo-labels from two aspects: the accuracy of predictions and the confidence of the predictions. For the first aspect, we propose a similarity graph structure learning (SGSL) model that considers the correlation between unlabeled and labeled samples, which facilitates the learning of more discriminative features and, thus, obtains more accurate predictions. For the second aspect, we propose an uncertainty-based graph convolutional network (UGCN), which can aggregate similar features based on the learned graph structure in the training phase, making the features more discriminative. It can also output the uncertainty of predictions in the pseudo-label generation phase, generating pseudo-labels only for unlabeled samples with low uncertainty; thus, reducing the noise in the pseudo-labels. Further, a positive and negative self-training framework is proposed, which combines the proposed SGSL model and UGCN into the self-training framework for end-to-end training. In addition, in order to introduce more supervised signals in the self-training process, negative pseudo-labels are generated for unlabeled samples with low prediction confidence, and then the positive and negative pseudo-labeled samples are trained together with a small number of labeled samples to improve the performance of semi-supervised learning. The code is available upon request. MDPI 2023-04-13 /pmc/articles/PMC10145307/ /pubmed/37112285 http://dx.doi.org/10.3390/s23083944 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wang, Yifan Huang, Yan Wang, Qicong Zhao, Chong Zhang, Zhenchang Chen, Jian Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning |
title | Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning |
title_full | Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning |
title_fullStr | Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning |
title_full_unstemmed | Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning |
title_short | Graph-Based Self-Training for Semi-Supervised Deep Similarity Learning |
title_sort | graph-based self-training for semi-supervised deep similarity learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10145307/ https://www.ncbi.nlm.nih.gov/pubmed/37112285 http://dx.doi.org/10.3390/s23083944 |
work_keys_str_mv | AT wangyifan graphbasedselftrainingforsemisuperviseddeepsimilaritylearning AT huangyan graphbasedselftrainingforsemisuperviseddeepsimilaritylearning AT wangqicong graphbasedselftrainingforsemisuperviseddeepsimilaritylearning AT zhaochong graphbasedselftrainingforsemisuperviseddeepsimilaritylearning AT zhangzhenchang graphbasedselftrainingforsemisuperviseddeepsimilaritylearning AT chenjian graphbasedselftrainingforsemisuperviseddeepsimilaritylearning |