Cargando…

A data-centric way to improve entity linking in knowledge-based question answering

Entity linking in knowledge-based question answering (KBQA) is intended to construct a mapping relation between a mention in a natural language question and an entity in the knowledge base. Most research in entity linking focuses on long text, but entity linking in open domain KBQA is more concerned...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Shuo, Zhou, Gang, Xia, Yi, Wu, Hao, Li, Zhufeng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10280402/
https://www.ncbi.nlm.nih.gov/pubmed/37346650
http://dx.doi.org/10.7717/peerj-cs.1233
_version_ 1785060786229477376
author Liu, Shuo
Zhou, Gang
Xia, Yi
Wu, Hao
Li, Zhufeng
author_facet Liu, Shuo
Zhou, Gang
Xia, Yi
Wu, Hao
Li, Zhufeng
author_sort Liu, Shuo
collection PubMed
description Entity linking in knowledge-based question answering (KBQA) is intended to construct a mapping relation between a mention in a natural language question and an entity in the knowledge base. Most research in entity linking focuses on long text, but entity linking in open domain KBQA is more concerned with short text. Many recent models have tried to extract the features of raw data by adjusting the neural network structure. However, the models only perform well with several datasets. We therefore concentrate on the data rather than the model itself and created a model DME (Domain information Mining and Explicit expressing) to extract domain information from short text and append it to the data. The entity linking model will be enhanced by training with DME-processed data. Besides, we also developed a novel negative sampling approach to make the model more robust. We conducted experiments using the large Chinese open source benchmark KgCLUE to assess model performance with DME-processed data. The experiments showed that our approach can improve entity linking in the baseline models without the need to change their structure and our approach is demonstrably transferable to other datasets.
format Online
Article
Text
id pubmed-10280402
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-102804022023-06-21 A data-centric way to improve entity linking in knowledge-based question answering Liu, Shuo Zhou, Gang Xia, Yi Wu, Hao Li, Zhufeng PeerJ Comput Sci Artificial Intelligence Entity linking in knowledge-based question answering (KBQA) is intended to construct a mapping relation between a mention in a natural language question and an entity in the knowledge base. Most research in entity linking focuses on long text, but entity linking in open domain KBQA is more concerned with short text. Many recent models have tried to extract the features of raw data by adjusting the neural network structure. However, the models only perform well with several datasets. We therefore concentrate on the data rather than the model itself and created a model DME (Domain information Mining and Explicit expressing) to extract domain information from short text and append it to the data. The entity linking model will be enhanced by training with DME-processed data. Besides, we also developed a novel negative sampling approach to make the model more robust. We conducted experiments using the large Chinese open source benchmark KgCLUE to assess model performance with DME-processed data. The experiments showed that our approach can improve entity linking in the baseline models without the need to change their structure and our approach is demonstrably transferable to other datasets. PeerJ Inc. 2023-02-09 /pmc/articles/PMC10280402/ /pubmed/37346650 http://dx.doi.org/10.7717/peerj-cs.1233 Text en ©2023 Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Artificial Intelligence
Liu, Shuo
Zhou, Gang
Xia, Yi
Wu, Hao
Li, Zhufeng
A data-centric way to improve entity linking in knowledge-based question answering
title A data-centric way to improve entity linking in knowledge-based question answering
title_full A data-centric way to improve entity linking in knowledge-based question answering
title_fullStr A data-centric way to improve entity linking in knowledge-based question answering
title_full_unstemmed A data-centric way to improve entity linking in knowledge-based question answering
title_short A data-centric way to improve entity linking in knowledge-based question answering
title_sort data-centric way to improve entity linking in knowledge-based question answering
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10280402/
https://www.ncbi.nlm.nih.gov/pubmed/37346650
http://dx.doi.org/10.7717/peerj-cs.1233
work_keys_str_mv AT liushuo adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT zhougang adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT xiayi adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT wuhao adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT lizhufeng adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT liushuo datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT zhougang datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT xiayi datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT wuhao datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering
AT lizhufeng datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering