Cargando…

Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning

Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the t...

Descripción completa

Detalles Bibliográficos
Autores principales: Sun, Weiwei, Liu, Shengquan, Liu, Yan, Kong, Lingqi, Jian, Zhaorui
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181062/
https://www.ncbi.nlm.nih.gov/pubmed/37177454
http://dx.doi.org/10.3390/s23094250
_version_ 1785041483663933440
author Sun, Weiwei
Liu, Shengquan
Liu, Yan
Kong, Lingqi
Jian, Zhaorui
author_facet Sun, Weiwei
Liu, Shengquan
Liu, Yan
Kong, Lingqi
Jian, Zhaorui
author_sort Sun, Weiwei
collection PubMed
description Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the text. Existing approaches typically use multi-layer encoders, such as Transformers, to generate hidden features of the source sequence. However, increasing the number of encoder layers can lead to the granularity of the representation becoming coarser and the hidden features of different words becoming more similar, potentially leading to the model’s misjudgment. To address this issue, a new method called the multi-granularity attention multi-scale self-learning network (MAML-NET) is proposed, which enhances the model’s understanding ability by utilizing different granularity representations of the source sequence. Additionally, MAML-NET can independently learn task-related information from both global and local dimensions based on the learned multi-granularity features through the proposed multi-scale self-learning attention mechanism. The experimental results on two information extraction tasks, named entity recognition and entity relationship extraction, demonstrated that the method was superior to the method based on machine reading comprehension and achieved the best performance on the five benchmark tests.
format Online
Article
Text
id pubmed-10181062
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-101810622023-05-13 Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning Sun, Weiwei Liu, Shengquan Liu, Yan Kong, Lingqi Jian, Zhaorui Sensors (Basel) Article Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the text. Existing approaches typically use multi-layer encoders, such as Transformers, to generate hidden features of the source sequence. However, increasing the number of encoder layers can lead to the granularity of the representation becoming coarser and the hidden features of different words becoming more similar, potentially leading to the model’s misjudgment. To address this issue, a new method called the multi-granularity attention multi-scale self-learning network (MAML-NET) is proposed, which enhances the model’s understanding ability by utilizing different granularity representations of the source sequence. Additionally, MAML-NET can independently learn task-related information from both global and local dimensions based on the learned multi-granularity features through the proposed multi-scale self-learning attention mechanism. The experimental results on two information extraction tasks, named entity recognition and entity relationship extraction, demonstrated that the method was superior to the method based on machine reading comprehension and achieved the best performance on the five benchmark tests. MDPI 2023-04-25 /pmc/articles/PMC10181062/ /pubmed/37177454 http://dx.doi.org/10.3390/s23094250 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sun, Weiwei
Liu, Shengquan
Liu, Yan
Kong, Lingqi
Jian, Zhaorui
Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
title Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
title_full Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
title_fullStr Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
title_full_unstemmed Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
title_short Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
title_sort information extraction network based on multi-granularity attention and multi-scale self-learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181062/
https://www.ncbi.nlm.nih.gov/pubmed/37177454
http://dx.doi.org/10.3390/s23094250
work_keys_str_mv AT sunweiwei informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning
AT liushengquan informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning
AT liuyan informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning
AT konglingqi informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning
AT jianzhaorui informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning