Cargando…

Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph

Recent years have witnessed the emergence of novel models for ad-hoc entity search in knowledge graphs of varying complexity. Since these models are based on direct term matching, their accuracy can suffer from a mismatch between vocabularies used in queries and entity descriptions. Although success...

Descripción completa

Detalles Bibliográficos
Autores principales: Nikolaev, Fedor, Kotov, Alexander
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148220/
http://dx.doi.org/10.1007/978-3-030-45439-5_10
_version_ 1783520546548350976
author Nikolaev, Fedor
Kotov, Alexander
author_facet Nikolaev, Fedor
Kotov, Alexander
author_sort Nikolaev, Fedor
collection PubMed
description Recent years have witnessed the emergence of novel models for ad-hoc entity search in knowledge graphs of varying complexity. Since these models are based on direct term matching, their accuracy can suffer from a mismatch between vocabularies used in queries and entity descriptions. Although successful applications of word embeddings and knowledge graph entity embeddings to address the issues of vocabulary mismatch in ad-hoc document retrieval and knowledge graph noisiness and incompleteness, respectively, have been reported in recent literature, the utility of joint word and entity embeddings for entity search in knowledge graphs has been relatively unexplored. In this paper, we propose Knowledge graph Entity and Word Embedding for Retrieval (KEWER), a novel method to embed entities and words into the same low-dimensional vector space, which takes into account a knowledge graph’s local structure and structural components, such as entities, attributes, and categories, and is designed specifically for entity search. KEWER is based on random walks over the knowledge graph and can be considered as a hybrid of word and network embedding methods. Similar to word embedding methods, KEWER utilizes contextual co-occurrences as training data, however, it treats words and entities as different objects. Similar to network embedding methods, KEWER takes into account knowledge graph’s local structure, however, it also differentiates between structural components. Experiments on publicly available entity search benchmarks and state-of-the-art word and joint word and entity embedding methods indicate that a combination of KEWER and BM25F results in a consistent improvement in retrieval accuracy over BM25F alone.
format Online
Article
Text
id pubmed-7148220
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-71482202020-04-13 Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph Nikolaev, Fedor Kotov, Alexander Advances in Information Retrieval Article Recent years have witnessed the emergence of novel models for ad-hoc entity search in knowledge graphs of varying complexity. Since these models are based on direct term matching, their accuracy can suffer from a mismatch between vocabularies used in queries and entity descriptions. Although successful applications of word embeddings and knowledge graph entity embeddings to address the issues of vocabulary mismatch in ad-hoc document retrieval and knowledge graph noisiness and incompleteness, respectively, have been reported in recent literature, the utility of joint word and entity embeddings for entity search in knowledge graphs has been relatively unexplored. In this paper, we propose Knowledge graph Entity and Word Embedding for Retrieval (KEWER), a novel method to embed entities and words into the same low-dimensional vector space, which takes into account a knowledge graph’s local structure and structural components, such as entities, attributes, and categories, and is designed specifically for entity search. KEWER is based on random walks over the knowledge graph and can be considered as a hybrid of word and network embedding methods. Similar to word embedding methods, KEWER utilizes contextual co-occurrences as training data, however, it treats words and entities as different objects. Similar to network embedding methods, KEWER takes into account knowledge graph’s local structure, however, it also differentiates between structural components. Experiments on publicly available entity search benchmarks and state-of-the-art word and joint word and entity embedding methods indicate that a combination of KEWER and BM25F results in a consistent improvement in retrieval accuracy over BM25F alone. 2020-03-17 /pmc/articles/PMC7148220/ http://dx.doi.org/10.1007/978-3-030-45439-5_10 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Nikolaev, Fedor
Kotov, Alexander
Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph
title Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph
title_full Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph
title_fullStr Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph
title_full_unstemmed Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph
title_short Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph
title_sort joint word and entity embeddings for entity retrieval from a knowledge graph
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148220/
http://dx.doi.org/10.1007/978-3-030-45439-5_10
work_keys_str_mv AT nikolaevfedor jointwordandentityembeddingsforentityretrievalfromaknowledgegraph
AT kotovalexander jointwordandentityembeddingsforentityretrievalfromaknowledgegraph