Cargando…
Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information
Knowledge graph completion can make knowledge graphs more complete, which is a meaningful research topic. However, the existing methods do not make full use of entity semantic information. Another challenge is that a deep model requires large-scale manually labelled data, which greatly increases man...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7597339/ https://www.ncbi.nlm.nih.gov/pubmed/33286937 http://dx.doi.org/10.3390/e22101168 |
_version_ | 1783602326164996096 |
---|---|
author | Zhang, Min Geng, Guohua Zeng, Sheng Jia, Huaping |
author_facet | Zhang, Min Geng, Guohua Zeng, Sheng Jia, Huaping |
author_sort | Zhang, Min |
collection | PubMed |
description | Knowledge graph completion can make knowledge graphs more complete, which is a meaningful research topic. However, the existing methods do not make full use of entity semantic information. Another challenge is that a deep model requires large-scale manually labelled data, which greatly increases manual labour. In order to alleviate the scarcity of labelled data in the field of cultural relics and capture the rich semantic information of entities, this paper proposes a model based on the Bidirectional Encoder Representations from Transformers (BERT) with entity-type information for the knowledge graph completion of the Chinese texts of cultural relics. In this work, the knowledge graph completion task is treated as a classification task, while the entities, relations and entity-type information are integrated as a textual sequence, and the Chinese characters are used as a token unit in which input representation is constructed by summing token, segment and position embeddings. A small number of labelled data are used to pre-train the model, and then, a large number of unlabelled data are used to fine-tune the pre-training model. The experiment results show that the BERT-KGC model with entity-type information can enrich the semantics information of the entities to reduce the degree of ambiguity of the entities and relations to some degree and achieve more effective performance than the baselines in triple classification, link prediction and relation prediction tasks using 35% of the labelled data of cultural relics. |
format | Online Article Text |
id | pubmed-7597339 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75973392020-11-09 Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information Zhang, Min Geng, Guohua Zeng, Sheng Jia, Huaping Entropy (Basel) Article Knowledge graph completion can make knowledge graphs more complete, which is a meaningful research topic. However, the existing methods do not make full use of entity semantic information. Another challenge is that a deep model requires large-scale manually labelled data, which greatly increases manual labour. In order to alleviate the scarcity of labelled data in the field of cultural relics and capture the rich semantic information of entities, this paper proposes a model based on the Bidirectional Encoder Representations from Transformers (BERT) with entity-type information for the knowledge graph completion of the Chinese texts of cultural relics. In this work, the knowledge graph completion task is treated as a classification task, while the entities, relations and entity-type information are integrated as a textual sequence, and the Chinese characters are used as a token unit in which input representation is constructed by summing token, segment and position embeddings. A small number of labelled data are used to pre-train the model, and then, a large number of unlabelled data are used to fine-tune the pre-training model. The experiment results show that the BERT-KGC model with entity-type information can enrich the semantics information of the entities to reduce the degree of ambiguity of the entities and relations to some degree and achieve more effective performance than the baselines in triple classification, link prediction and relation prediction tasks using 35% of the labelled data of cultural relics. MDPI 2020-10-16 /pmc/articles/PMC7597339/ /pubmed/33286937 http://dx.doi.org/10.3390/e22101168 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhang, Min Geng, Guohua Zeng, Sheng Jia, Huaping Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information |
title | Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information |
title_full | Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information |
title_fullStr | Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information |
title_full_unstemmed | Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information |
title_short | Knowledge Graph Completion for the Chinese Text of Cultural Relics Based on Bidirectional Encoder Representations from Transformers with Entity-Type Information |
title_sort | knowledge graph completion for the chinese text of cultural relics based on bidirectional encoder representations from transformers with entity-type information |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7597339/ https://www.ncbi.nlm.nih.gov/pubmed/33286937 http://dx.doi.org/10.3390/e22101168 |
work_keys_str_mv | AT zhangmin knowledgegraphcompletionforthechinesetextofculturalrelicsbasedonbidirectionalencoderrepresentationsfromtransformerswithentitytypeinformation AT gengguohua knowledgegraphcompletionforthechinesetextofculturalrelicsbasedonbidirectionalencoderrepresentationsfromtransformerswithentitytypeinformation AT zengsheng knowledgegraphcompletionforthechinesetextofculturalrelicsbasedonbidirectionalencoderrepresentationsfromtransformerswithentitytypeinformation AT jiahuaping knowledgegraphcompletionforthechinesetextofculturalrelicsbasedonbidirectionalencoderrepresentationsfromtransformerswithentitytypeinformation |