Cargando…

Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model

The purpose of this study is to solve the effective way of domain-specific knowledge graph construction from information to knowledge. We propose the deep learning algorithm to extract entities and relationship from open-source intelligence by the RoBERTa-wwm-ext pretraining model and a knowledge fu...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Xingli, Zhao, Wei, Ma, Haiqun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9581622/
https://www.ncbi.nlm.nih.gov/pubmed/36275948
http://dx.doi.org/10.1155/2022/8656013
_version_ 1784812665677283328
author Liu, Xingli
Zhao, Wei
Ma, Haiqun
author_facet Liu, Xingli
Zhao, Wei
Ma, Haiqun
author_sort Liu, Xingli
collection PubMed
description The purpose of this study is to solve the effective way of domain-specific knowledge graph construction from information to knowledge. We propose the deep learning algorithm to extract entities and relationship from open-source intelligence by the RoBERTa-wwm-ext pretraining model and a knowledge fusion framework based on the longest common attribute entity alignment technology and bring in different text similarity algorithms and classification algorithms for verification. The experimental research showed that the named entity recognition model using the RoBERTa-wwm-ext pretrained model achieves the best results in terms of recall rate and F1 value, first, and the F value of RoBERTa-wwm-ext + BiLSTM + CRF reached up to 83.07%. Second, the RoBERTa-wwm-ext relationship extraction model has achieved the best results; compared with the relation extraction model based on recurrent neural network, it is improved by about 20%∼30%. Finally, the entity alignment algorithm based on the attribute similarity of the longest common subsequence has achieved the best results on the whole. The findings of this study provide an effective way to complete knowledge graph construction in domain-specific texts. The research serves as a first step for future research, for example, domain-specific intelligent Q&A.
format Online
Article
Text
id pubmed-9581622
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-95816222022-10-20 Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model Liu, Xingli Zhao, Wei Ma, Haiqun Comput Intell Neurosci Research Article The purpose of this study is to solve the effective way of domain-specific knowledge graph construction from information to knowledge. We propose the deep learning algorithm to extract entities and relationship from open-source intelligence by the RoBERTa-wwm-ext pretraining model and a knowledge fusion framework based on the longest common attribute entity alignment technology and bring in different text similarity algorithms and classification algorithms for verification. The experimental research showed that the named entity recognition model using the RoBERTa-wwm-ext pretrained model achieves the best results in terms of recall rate and F1 value, first, and the F value of RoBERTa-wwm-ext + BiLSTM + CRF reached up to 83.07%. Second, the RoBERTa-wwm-ext relationship extraction model has achieved the best results; compared with the relation extraction model based on recurrent neural network, it is improved by about 20%∼30%. Finally, the entity alignment algorithm based on the attribute similarity of the longest common subsequence has achieved the best results on the whole. The findings of this study provide an effective way to complete knowledge graph construction in domain-specific texts. The research serves as a first step for future research, for example, domain-specific intelligent Q&A. Hindawi 2022-10-12 /pmc/articles/PMC9581622/ /pubmed/36275948 http://dx.doi.org/10.1155/2022/8656013 Text en Copyright © 2022 Xingli Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Liu, Xingli
Zhao, Wei
Ma, Haiqun
Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
title Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
title_full Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
title_fullStr Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
title_full_unstemmed Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
title_short Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
title_sort research on domain-specific knowledge graph based on the roberta-wwm-ext pretraining model
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9581622/
https://www.ncbi.nlm.nih.gov/pubmed/36275948
http://dx.doi.org/10.1155/2022/8656013
work_keys_str_mv AT liuxingli researchondomainspecificknowledgegraphbasedontherobertawwmextpretrainingmodel
AT zhaowei researchondomainspecificknowledgegraphbasedontherobertawwmextpretrainingmodel
AT mahaiqun researchondomainspecificknowledgegraphbasedontherobertawwmextpretrainingmodel