Cargando…

A multi-view contrastive learning for heterogeneous network embedding

Graph contrastive learning has been developed to learn discriminative node representations on homogeneous graphs. However, it is not clear how to augment the heterogeneous graphs without substantially altering the underlying semantics or how to design appropriate pretext tasks to fully capture the r...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Qi, Chen, Wenping, Fang, Zhaoxi, Ying, Changtian, Wang, Chen
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10130187/
https://www.ncbi.nlm.nih.gov/pubmed/37185784
http://dx.doi.org/10.1038/s41598-023-33324-7
_version_ 1785030915497394176
author Li, Qi
Chen, Wenping
Fang, Zhaoxi
Ying, Changtian
Wang, Chen
author_facet Li, Qi
Chen, Wenping
Fang, Zhaoxi
Ying, Changtian
Wang, Chen
author_sort Li, Qi
collection PubMed
description Graph contrastive learning has been developed to learn discriminative node representations on homogeneous graphs. However, it is not clear how to augment the heterogeneous graphs without substantially altering the underlying semantics or how to design appropriate pretext tasks to fully capture the rich semantics preserved in heterogeneous information networks (HINs). Moreover, early investigations demonstrate that contrastive learning suffer from sampling bias, whereas conventional debiasing techniques (e.g., hard negative mining) are empirically shown to be inadequate for graph contrastive learning. How to mitigate the sampling bias on heterogeneous graphs is another important yet neglected problem. To address the aforementioned challenges, we propose a novel multi-view heterogeneous graph contrastive learning framework in this paper. We use metapaths, each of which depicts a complementary element of HINs, as the augmentation to generate multiple subgraphs (i.e., multi-views), and propose a novel pretext task to maximize the coherence between each pair of metapath-induced views. Furthermore, we employ a positive sampling strategy to explicitly select hard positives by jointly considering semantics and structures preserved on each metapath view to alleviate the sampling bias. Extensive experiments demonstrate MCL consistently outperforms state-of-the-art baselines on five real-world benchmark datasets and even its supervised counterparts in some settings.
format Online
Article
Text
id pubmed-10130187
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-101301872023-04-27 A multi-view contrastive learning for heterogeneous network embedding Li, Qi Chen, Wenping Fang, Zhaoxi Ying, Changtian Wang, Chen Sci Rep Article Graph contrastive learning has been developed to learn discriminative node representations on homogeneous graphs. However, it is not clear how to augment the heterogeneous graphs without substantially altering the underlying semantics or how to design appropriate pretext tasks to fully capture the rich semantics preserved in heterogeneous information networks (HINs). Moreover, early investigations demonstrate that contrastive learning suffer from sampling bias, whereas conventional debiasing techniques (e.g., hard negative mining) are empirically shown to be inadequate for graph contrastive learning. How to mitigate the sampling bias on heterogeneous graphs is another important yet neglected problem. To address the aforementioned challenges, we propose a novel multi-view heterogeneous graph contrastive learning framework in this paper. We use metapaths, each of which depicts a complementary element of HINs, as the augmentation to generate multiple subgraphs (i.e., multi-views), and propose a novel pretext task to maximize the coherence between each pair of metapath-induced views. Furthermore, we employ a positive sampling strategy to explicitly select hard positives by jointly considering semantics and structures preserved on each metapath view to alleviate the sampling bias. Extensive experiments demonstrate MCL consistently outperforms state-of-the-art baselines on five real-world benchmark datasets and even its supervised counterparts in some settings. Nature Publishing Group UK 2023-04-25 /pmc/articles/PMC10130187/ /pubmed/37185784 http://dx.doi.org/10.1038/s41598-023-33324-7 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Li, Qi
Chen, Wenping
Fang, Zhaoxi
Ying, Changtian
Wang, Chen
A multi-view contrastive learning for heterogeneous network embedding
title A multi-view contrastive learning for heterogeneous network embedding
title_full A multi-view contrastive learning for heterogeneous network embedding
title_fullStr A multi-view contrastive learning for heterogeneous network embedding
title_full_unstemmed A multi-view contrastive learning for heterogeneous network embedding
title_short A multi-view contrastive learning for heterogeneous network embedding
title_sort multi-view contrastive learning for heterogeneous network embedding
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10130187/
https://www.ncbi.nlm.nih.gov/pubmed/37185784
http://dx.doi.org/10.1038/s41598-023-33324-7
work_keys_str_mv AT liqi amultiviewcontrastivelearningforheterogeneousnetworkembedding
AT chenwenping amultiviewcontrastivelearningforheterogeneousnetworkembedding
AT fangzhaoxi amultiviewcontrastivelearningforheterogeneousnetworkembedding
AT yingchangtian amultiviewcontrastivelearningforheterogeneousnetworkembedding
AT wangchen amultiviewcontrastivelearningforheterogeneousnetworkembedding
AT liqi multiviewcontrastivelearningforheterogeneousnetworkembedding
AT chenwenping multiviewcontrastivelearningforheterogeneousnetworkembedding
AT fangzhaoxi multiviewcontrastivelearningforheterogeneousnetworkembedding
AT yingchangtian multiviewcontrastivelearningforheterogeneousnetworkembedding
AT wangchen multiviewcontrastivelearningforheterogeneousnetworkembedding