Cargando…
Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model
The purpose of this study is to solve the effective way of domain-specific knowledge graph construction from information to knowledge. We propose the deep learning algorithm to extract entities and relationship from open-source intelligence by the RoBERTa-wwm-ext pretraining model and a knowledge fu...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9581622/ https://www.ncbi.nlm.nih.gov/pubmed/36275948 http://dx.doi.org/10.1155/2022/8656013 |
_version_ | 1784812665677283328 |
---|---|
author | Liu, Xingli Zhao, Wei Ma, Haiqun |
author_facet | Liu, Xingli Zhao, Wei Ma, Haiqun |
author_sort | Liu, Xingli |
collection | PubMed |
description | The purpose of this study is to solve the effective way of domain-specific knowledge graph construction from information to knowledge. We propose the deep learning algorithm to extract entities and relationship from open-source intelligence by the RoBERTa-wwm-ext pretraining model and a knowledge fusion framework based on the longest common attribute entity alignment technology and bring in different text similarity algorithms and classification algorithms for verification. The experimental research showed that the named entity recognition model using the RoBERTa-wwm-ext pretrained model achieves the best results in terms of recall rate and F1 value, first, and the F value of RoBERTa-wwm-ext + BiLSTM + CRF reached up to 83.07%. Second, the RoBERTa-wwm-ext relationship extraction model has achieved the best results; compared with the relation extraction model based on recurrent neural network, it is improved by about 20%∼30%. Finally, the entity alignment algorithm based on the attribute similarity of the longest common subsequence has achieved the best results on the whole. The findings of this study provide an effective way to complete knowledge graph construction in domain-specific texts. The research serves as a first step for future research, for example, domain-specific intelligent Q&A. |
format | Online Article Text |
id | pubmed-9581622 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-95816222022-10-20 Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model Liu, Xingli Zhao, Wei Ma, Haiqun Comput Intell Neurosci Research Article The purpose of this study is to solve the effective way of domain-specific knowledge graph construction from information to knowledge. We propose the deep learning algorithm to extract entities and relationship from open-source intelligence by the RoBERTa-wwm-ext pretraining model and a knowledge fusion framework based on the longest common attribute entity alignment technology and bring in different text similarity algorithms and classification algorithms for verification. The experimental research showed that the named entity recognition model using the RoBERTa-wwm-ext pretrained model achieves the best results in terms of recall rate and F1 value, first, and the F value of RoBERTa-wwm-ext + BiLSTM + CRF reached up to 83.07%. Second, the RoBERTa-wwm-ext relationship extraction model has achieved the best results; compared with the relation extraction model based on recurrent neural network, it is improved by about 20%∼30%. Finally, the entity alignment algorithm based on the attribute similarity of the longest common subsequence has achieved the best results on the whole. The findings of this study provide an effective way to complete knowledge graph construction in domain-specific texts. The research serves as a first step for future research, for example, domain-specific intelligent Q&A. Hindawi 2022-10-12 /pmc/articles/PMC9581622/ /pubmed/36275948 http://dx.doi.org/10.1155/2022/8656013 Text en Copyright © 2022 Xingli Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Liu, Xingli Zhao, Wei Ma, Haiqun Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model |
title | Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model |
title_full | Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model |
title_fullStr | Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model |
title_full_unstemmed | Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model |
title_short | Research on Domain-Specific Knowledge Graph Based on the RoBERTa-wwm-ext Pretraining Model |
title_sort | research on domain-specific knowledge graph based on the roberta-wwm-ext pretraining model |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9581622/ https://www.ncbi.nlm.nih.gov/pubmed/36275948 http://dx.doi.org/10.1155/2022/8656013 |
work_keys_str_mv | AT liuxingli researchondomainspecificknowledgegraphbasedontherobertawwmextpretrainingmodel AT zhaowei researchondomainspecificknowledgegraphbasedontherobertawwmextpretrainingmodel AT mahaiqun researchondomainspecificknowledgegraphbasedontherobertawwmextpretrainingmodel |