Cargando…
A data-centric way to improve entity linking in knowledge-based question answering
Entity linking in knowledge-based question answering (KBQA) is intended to construct a mapping relation between a mention in a natural language question and an entity in the knowledge base. Most research in entity linking focuses on long text, but entity linking in open domain KBQA is more concerned...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
PeerJ Inc.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10280402/ https://www.ncbi.nlm.nih.gov/pubmed/37346650 http://dx.doi.org/10.7717/peerj-cs.1233 |
_version_ | 1785060786229477376 |
---|---|
author | Liu, Shuo Zhou, Gang Xia, Yi Wu, Hao Li, Zhufeng |
author_facet | Liu, Shuo Zhou, Gang Xia, Yi Wu, Hao Li, Zhufeng |
author_sort | Liu, Shuo |
collection | PubMed |
description | Entity linking in knowledge-based question answering (KBQA) is intended to construct a mapping relation between a mention in a natural language question and an entity in the knowledge base. Most research in entity linking focuses on long text, but entity linking in open domain KBQA is more concerned with short text. Many recent models have tried to extract the features of raw data by adjusting the neural network structure. However, the models only perform well with several datasets. We therefore concentrate on the data rather than the model itself and created a model DME (Domain information Mining and Explicit expressing) to extract domain information from short text and append it to the data. The entity linking model will be enhanced by training with DME-processed data. Besides, we also developed a novel negative sampling approach to make the model more robust. We conducted experiments using the large Chinese open source benchmark KgCLUE to assess model performance with DME-processed data. The experiments showed that our approach can improve entity linking in the baseline models without the need to change their structure and our approach is demonstrably transferable to other datasets. |
format | Online Article Text |
id | pubmed-10280402 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | PeerJ Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-102804022023-06-21 A data-centric way to improve entity linking in knowledge-based question answering Liu, Shuo Zhou, Gang Xia, Yi Wu, Hao Li, Zhufeng PeerJ Comput Sci Artificial Intelligence Entity linking in knowledge-based question answering (KBQA) is intended to construct a mapping relation between a mention in a natural language question and an entity in the knowledge base. Most research in entity linking focuses on long text, but entity linking in open domain KBQA is more concerned with short text. Many recent models have tried to extract the features of raw data by adjusting the neural network structure. However, the models only perform well with several datasets. We therefore concentrate on the data rather than the model itself and created a model DME (Domain information Mining and Explicit expressing) to extract domain information from short text and append it to the data. The entity linking model will be enhanced by training with DME-processed data. Besides, we also developed a novel negative sampling approach to make the model more robust. We conducted experiments using the large Chinese open source benchmark KgCLUE to assess model performance with DME-processed data. The experiments showed that our approach can improve entity linking in the baseline models without the need to change their structure and our approach is demonstrably transferable to other datasets. PeerJ Inc. 2023-02-09 /pmc/articles/PMC10280402/ /pubmed/37346650 http://dx.doi.org/10.7717/peerj-cs.1233 Text en ©2023 Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
spellingShingle | Artificial Intelligence Liu, Shuo Zhou, Gang Xia, Yi Wu, Hao Li, Zhufeng A data-centric way to improve entity linking in knowledge-based question answering |
title | A data-centric way to improve entity linking in knowledge-based question answering |
title_full | A data-centric way to improve entity linking in knowledge-based question answering |
title_fullStr | A data-centric way to improve entity linking in knowledge-based question answering |
title_full_unstemmed | A data-centric way to improve entity linking in knowledge-based question answering |
title_short | A data-centric way to improve entity linking in knowledge-based question answering |
title_sort | data-centric way to improve entity linking in knowledge-based question answering |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10280402/ https://www.ncbi.nlm.nih.gov/pubmed/37346650 http://dx.doi.org/10.7717/peerj-cs.1233 |
work_keys_str_mv | AT liushuo adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT zhougang adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT xiayi adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT wuhao adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT lizhufeng adatacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT liushuo datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT zhougang datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT xiayi datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT wuhao datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering AT lizhufeng datacentricwaytoimproveentitylinkinginknowledgebasedquestionanswering |