Cargando…
A multi-view contrastive learning for heterogeneous network embedding
Graph contrastive learning has been developed to learn discriminative node representations on homogeneous graphs. However, it is not clear how to augment the heterogeneous graphs without substantially altering the underlying semantics or how to design appropriate pretext tasks to fully capture the r...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10130187/ https://www.ncbi.nlm.nih.gov/pubmed/37185784 http://dx.doi.org/10.1038/s41598-023-33324-7 |
_version_ | 1785030915497394176 |
---|---|
author | Li, Qi Chen, Wenping Fang, Zhaoxi Ying, Changtian Wang, Chen |
author_facet | Li, Qi Chen, Wenping Fang, Zhaoxi Ying, Changtian Wang, Chen |
author_sort | Li, Qi |
collection | PubMed |
description | Graph contrastive learning has been developed to learn discriminative node representations on homogeneous graphs. However, it is not clear how to augment the heterogeneous graphs without substantially altering the underlying semantics or how to design appropriate pretext tasks to fully capture the rich semantics preserved in heterogeneous information networks (HINs). Moreover, early investigations demonstrate that contrastive learning suffer from sampling bias, whereas conventional debiasing techniques (e.g., hard negative mining) are empirically shown to be inadequate for graph contrastive learning. How to mitigate the sampling bias on heterogeneous graphs is another important yet neglected problem. To address the aforementioned challenges, we propose a novel multi-view heterogeneous graph contrastive learning framework in this paper. We use metapaths, each of which depicts a complementary element of HINs, as the augmentation to generate multiple subgraphs (i.e., multi-views), and propose a novel pretext task to maximize the coherence between each pair of metapath-induced views. Furthermore, we employ a positive sampling strategy to explicitly select hard positives by jointly considering semantics and structures preserved on each metapath view to alleviate the sampling bias. Extensive experiments demonstrate MCL consistently outperforms state-of-the-art baselines on five real-world benchmark datasets and even its supervised counterparts in some settings. |
format | Online Article Text |
id | pubmed-10130187 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-101301872023-04-27 A multi-view contrastive learning for heterogeneous network embedding Li, Qi Chen, Wenping Fang, Zhaoxi Ying, Changtian Wang, Chen Sci Rep Article Graph contrastive learning has been developed to learn discriminative node representations on homogeneous graphs. However, it is not clear how to augment the heterogeneous graphs without substantially altering the underlying semantics or how to design appropriate pretext tasks to fully capture the rich semantics preserved in heterogeneous information networks (HINs). Moreover, early investigations demonstrate that contrastive learning suffer from sampling bias, whereas conventional debiasing techniques (e.g., hard negative mining) are empirically shown to be inadequate for graph contrastive learning. How to mitigate the sampling bias on heterogeneous graphs is another important yet neglected problem. To address the aforementioned challenges, we propose a novel multi-view heterogeneous graph contrastive learning framework in this paper. We use metapaths, each of which depicts a complementary element of HINs, as the augmentation to generate multiple subgraphs (i.e., multi-views), and propose a novel pretext task to maximize the coherence between each pair of metapath-induced views. Furthermore, we employ a positive sampling strategy to explicitly select hard positives by jointly considering semantics and structures preserved on each metapath view to alleviate the sampling bias. Extensive experiments demonstrate MCL consistently outperforms state-of-the-art baselines on five real-world benchmark datasets and even its supervised counterparts in some settings. Nature Publishing Group UK 2023-04-25 /pmc/articles/PMC10130187/ /pubmed/37185784 http://dx.doi.org/10.1038/s41598-023-33324-7 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Li, Qi Chen, Wenping Fang, Zhaoxi Ying, Changtian Wang, Chen A multi-view contrastive learning for heterogeneous network embedding |
title | A multi-view contrastive learning for heterogeneous network embedding |
title_full | A multi-view contrastive learning for heterogeneous network embedding |
title_fullStr | A multi-view contrastive learning for heterogeneous network embedding |
title_full_unstemmed | A multi-view contrastive learning for heterogeneous network embedding |
title_short | A multi-view contrastive learning for heterogeneous network embedding |
title_sort | multi-view contrastive learning for heterogeneous network embedding |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10130187/ https://www.ncbi.nlm.nih.gov/pubmed/37185784 http://dx.doi.org/10.1038/s41598-023-33324-7 |
work_keys_str_mv | AT liqi amultiviewcontrastivelearningforheterogeneousnetworkembedding AT chenwenping amultiviewcontrastivelearningforheterogeneousnetworkembedding AT fangzhaoxi amultiviewcontrastivelearningforheterogeneousnetworkembedding AT yingchangtian amultiviewcontrastivelearningforheterogeneousnetworkembedding AT wangchen amultiviewcontrastivelearningforheterogeneousnetworkembedding AT liqi multiviewcontrastivelearningforheterogeneousnetworkembedding AT chenwenping multiviewcontrastivelearningforheterogeneousnetworkembedding AT fangzhaoxi multiviewcontrastivelearningforheterogeneousnetworkembedding AT yingchangtian multiviewcontrastivelearningforheterogeneousnetworkembedding AT wangchen multiviewcontrastivelearningforheterogeneousnetworkembedding |