Cargando…

Query-Specific Deep Embedding of Content-Rich Network

In this paper, we propose to embed a content-rich network for the purpose of similarity searching for a query node. In this network, besides the information of the nodes and edges, we also have the content of each node. We use the convolutional neural network (CNN) to represent the content of each n...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Yue, Wang, Hongqi, Yu, Liqun, Cooper, Sarah Yvonne, Wang, Jing-Yan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7468613/
https://www.ncbi.nlm.nih.gov/pubmed/32908476
http://dx.doi.org/10.1155/2020/5943798
_version_ 1783578255775760384
author Li, Yue
Wang, Hongqi
Yu, Liqun
Cooper, Sarah Yvonne
Wang, Jing-Yan
author_facet Li, Yue
Wang, Hongqi
Yu, Liqun
Cooper, Sarah Yvonne
Wang, Jing-Yan
author_sort Li, Yue
collection PubMed
description In this paper, we propose to embed a content-rich network for the purpose of similarity searching for a query node. In this network, besides the information of the nodes and edges, we also have the content of each node. We use the convolutional neural network (CNN) to represent the content of each node and then use the graph convolutional network (GCN) to further represent the node by merging the representations of its neighboring nodes. The GCN output is further fed to a deep encoder-decoder model to convert each node to a Gaussian distribution and then convert back to its node identity. The dissimilarity between the two nodes is measured by the Wasserstein distance between their Gaussian distributions. We define the nodes of the network to be positives if they are relevant to the query node and negative if they are irrelevant. The labeling of the positives/negatives is based on an upper bound and a lower bound of the Wasserstein distances between the candidate nodes and the query nodes. We learn the parameters of CNN, GCN, encoder-decoder model, Gaussian distributions, and the upper bound and lower bounds jointly. The learning problem is modeled as a minimization problem to minimize the losses of node identification, network structure preservation, positive/negative query-specific relevance-guild distance, and model complexity. An iterative algorithm is developed to solve the minimization problem. We conducted experiments over benchmark networks, especially innovation networks, to verify the effectiveness of the proposed method and showed its advantage over the state-of-the-art methods.
format Online
Article
Text
id pubmed-7468613
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-74686132020-09-08 Query-Specific Deep Embedding of Content-Rich Network Li, Yue Wang, Hongqi Yu, Liqun Cooper, Sarah Yvonne Wang, Jing-Yan Comput Intell Neurosci Research Article In this paper, we propose to embed a content-rich network for the purpose of similarity searching for a query node. In this network, besides the information of the nodes and edges, we also have the content of each node. We use the convolutional neural network (CNN) to represent the content of each node and then use the graph convolutional network (GCN) to further represent the node by merging the representations of its neighboring nodes. The GCN output is further fed to a deep encoder-decoder model to convert each node to a Gaussian distribution and then convert back to its node identity. The dissimilarity between the two nodes is measured by the Wasserstein distance between their Gaussian distributions. We define the nodes of the network to be positives if they are relevant to the query node and negative if they are irrelevant. The labeling of the positives/negatives is based on an upper bound and a lower bound of the Wasserstein distances between the candidate nodes and the query nodes. We learn the parameters of CNN, GCN, encoder-decoder model, Gaussian distributions, and the upper bound and lower bounds jointly. The learning problem is modeled as a minimization problem to minimize the losses of node identification, network structure preservation, positive/negative query-specific relevance-guild distance, and model complexity. An iterative algorithm is developed to solve the minimization problem. We conducted experiments over benchmark networks, especially innovation networks, to verify the effectiveness of the proposed method and showed its advantage over the state-of-the-art methods. Hindawi 2020-08-25 /pmc/articles/PMC7468613/ /pubmed/32908476 http://dx.doi.org/10.1155/2020/5943798 Text en Copyright © 2020 Yue Li et al. http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Li, Yue
Wang, Hongqi
Yu, Liqun
Cooper, Sarah Yvonne
Wang, Jing-Yan
Query-Specific Deep Embedding of Content-Rich Network
title Query-Specific Deep Embedding of Content-Rich Network
title_full Query-Specific Deep Embedding of Content-Rich Network
title_fullStr Query-Specific Deep Embedding of Content-Rich Network
title_full_unstemmed Query-Specific Deep Embedding of Content-Rich Network
title_short Query-Specific Deep Embedding of Content-Rich Network
title_sort query-specific deep embedding of content-rich network
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7468613/
https://www.ncbi.nlm.nih.gov/pubmed/32908476
http://dx.doi.org/10.1155/2020/5943798
work_keys_str_mv AT liyue queryspecificdeepembeddingofcontentrichnetwork
AT wanghongqi queryspecificdeepembeddingofcontentrichnetwork
AT yuliqun queryspecificdeepembeddingofcontentrichnetwork
AT coopersarahyvonne queryspecificdeepembeddingofcontentrichnetwork
AT wangjingyan queryspecificdeepembeddingofcontentrichnetwork