Cargando…
Query-Specific Deep Embedding of Content-Rich Network
In this paper, we propose to embed a content-rich network for the purpose of similarity searching for a query node. In this network, besides the information of the nodes and edges, we also have the content of each node. We use the convolutional neural network (CNN) to represent the content of each n...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7468613/ https://www.ncbi.nlm.nih.gov/pubmed/32908476 http://dx.doi.org/10.1155/2020/5943798 |
_version_ | 1783578255775760384 |
---|---|
author | Li, Yue Wang, Hongqi Yu, Liqun Cooper, Sarah Yvonne Wang, Jing-Yan |
author_facet | Li, Yue Wang, Hongqi Yu, Liqun Cooper, Sarah Yvonne Wang, Jing-Yan |
author_sort | Li, Yue |
collection | PubMed |
description | In this paper, we propose to embed a content-rich network for the purpose of similarity searching for a query node. In this network, besides the information of the nodes and edges, we also have the content of each node. We use the convolutional neural network (CNN) to represent the content of each node and then use the graph convolutional network (GCN) to further represent the node by merging the representations of its neighboring nodes. The GCN output is further fed to a deep encoder-decoder model to convert each node to a Gaussian distribution and then convert back to its node identity. The dissimilarity between the two nodes is measured by the Wasserstein distance between their Gaussian distributions. We define the nodes of the network to be positives if they are relevant to the query node and negative if they are irrelevant. The labeling of the positives/negatives is based on an upper bound and a lower bound of the Wasserstein distances between the candidate nodes and the query nodes. We learn the parameters of CNN, GCN, encoder-decoder model, Gaussian distributions, and the upper bound and lower bounds jointly. The learning problem is modeled as a minimization problem to minimize the losses of node identification, network structure preservation, positive/negative query-specific relevance-guild distance, and model complexity. An iterative algorithm is developed to solve the minimization problem. We conducted experiments over benchmark networks, especially innovation networks, to verify the effectiveness of the proposed method and showed its advantage over the state-of-the-art methods. |
format | Online Article Text |
id | pubmed-7468613 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-74686132020-09-08 Query-Specific Deep Embedding of Content-Rich Network Li, Yue Wang, Hongqi Yu, Liqun Cooper, Sarah Yvonne Wang, Jing-Yan Comput Intell Neurosci Research Article In this paper, we propose to embed a content-rich network for the purpose of similarity searching for a query node. In this network, besides the information of the nodes and edges, we also have the content of each node. We use the convolutional neural network (CNN) to represent the content of each node and then use the graph convolutional network (GCN) to further represent the node by merging the representations of its neighboring nodes. The GCN output is further fed to a deep encoder-decoder model to convert each node to a Gaussian distribution and then convert back to its node identity. The dissimilarity between the two nodes is measured by the Wasserstein distance between their Gaussian distributions. We define the nodes of the network to be positives if they are relevant to the query node and negative if they are irrelevant. The labeling of the positives/negatives is based on an upper bound and a lower bound of the Wasserstein distances between the candidate nodes and the query nodes. We learn the parameters of CNN, GCN, encoder-decoder model, Gaussian distributions, and the upper bound and lower bounds jointly. The learning problem is modeled as a minimization problem to minimize the losses of node identification, network structure preservation, positive/negative query-specific relevance-guild distance, and model complexity. An iterative algorithm is developed to solve the minimization problem. We conducted experiments over benchmark networks, especially innovation networks, to verify the effectiveness of the proposed method and showed its advantage over the state-of-the-art methods. Hindawi 2020-08-25 /pmc/articles/PMC7468613/ /pubmed/32908476 http://dx.doi.org/10.1155/2020/5943798 Text en Copyright © 2020 Yue Li et al. http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Li, Yue Wang, Hongqi Yu, Liqun Cooper, Sarah Yvonne Wang, Jing-Yan Query-Specific Deep Embedding of Content-Rich Network |
title | Query-Specific Deep Embedding of Content-Rich Network |
title_full | Query-Specific Deep Embedding of Content-Rich Network |
title_fullStr | Query-Specific Deep Embedding of Content-Rich Network |
title_full_unstemmed | Query-Specific Deep Embedding of Content-Rich Network |
title_short | Query-Specific Deep Embedding of Content-Rich Network |
title_sort | query-specific deep embedding of content-rich network |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7468613/ https://www.ncbi.nlm.nih.gov/pubmed/32908476 http://dx.doi.org/10.1155/2020/5943798 |
work_keys_str_mv | AT liyue queryspecificdeepembeddingofcontentrichnetwork AT wanghongqi queryspecificdeepembeddingofcontentrichnetwork AT yuliqun queryspecificdeepembeddingofcontentrichnetwork AT coopersarahyvonne queryspecificdeepembeddingofcontentrichnetwork AT wangjingyan queryspecificdeepembeddingofcontentrichnetwork |