Cargando…
Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the t...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181062/ https://www.ncbi.nlm.nih.gov/pubmed/37177454 http://dx.doi.org/10.3390/s23094250 |
_version_ | 1785041483663933440 |
---|---|
author | Sun, Weiwei Liu, Shengquan Liu, Yan Kong, Lingqi Jian, Zhaorui |
author_facet | Sun, Weiwei Liu, Shengquan Liu, Yan Kong, Lingqi Jian, Zhaorui |
author_sort | Sun, Weiwei |
collection | PubMed |
description | Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the text. Existing approaches typically use multi-layer encoders, such as Transformers, to generate hidden features of the source sequence. However, increasing the number of encoder layers can lead to the granularity of the representation becoming coarser and the hidden features of different words becoming more similar, potentially leading to the model’s misjudgment. To address this issue, a new method called the multi-granularity attention multi-scale self-learning network (MAML-NET) is proposed, which enhances the model’s understanding ability by utilizing different granularity representations of the source sequence. Additionally, MAML-NET can independently learn task-related information from both global and local dimensions based on the learned multi-granularity features through the proposed multi-scale self-learning attention mechanism. The experimental results on two information extraction tasks, named entity recognition and entity relationship extraction, demonstrated that the method was superior to the method based on machine reading comprehension and achieved the best performance on the five benchmark tests. |
format | Online Article Text |
id | pubmed-10181062 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-101810622023-05-13 Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning Sun, Weiwei Liu, Shengquan Liu, Yan Kong, Lingqi Jian, Zhaorui Sensors (Basel) Article Transforming the task of information extraction into a machine reading comprehension (MRC) framework has shown promising results. The MRC model takes the context and query as the inputs to the encoder, and the decoder extracts one or more text spans as answers (entities and relationships) from the text. Existing approaches typically use multi-layer encoders, such as Transformers, to generate hidden features of the source sequence. However, increasing the number of encoder layers can lead to the granularity of the representation becoming coarser and the hidden features of different words becoming more similar, potentially leading to the model’s misjudgment. To address this issue, a new method called the multi-granularity attention multi-scale self-learning network (MAML-NET) is proposed, which enhances the model’s understanding ability by utilizing different granularity representations of the source sequence. Additionally, MAML-NET can independently learn task-related information from both global and local dimensions based on the learned multi-granularity features through the proposed multi-scale self-learning attention mechanism. The experimental results on two information extraction tasks, named entity recognition and entity relationship extraction, demonstrated that the method was superior to the method based on machine reading comprehension and achieved the best performance on the five benchmark tests. MDPI 2023-04-25 /pmc/articles/PMC10181062/ /pubmed/37177454 http://dx.doi.org/10.3390/s23094250 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Sun, Weiwei Liu, Shengquan Liu, Yan Kong, Lingqi Jian, Zhaorui Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning |
title | Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning |
title_full | Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning |
title_fullStr | Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning |
title_full_unstemmed | Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning |
title_short | Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning |
title_sort | information extraction network based on multi-granularity attention and multi-scale self-learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10181062/ https://www.ncbi.nlm.nih.gov/pubmed/37177454 http://dx.doi.org/10.3390/s23094250 |
work_keys_str_mv | AT sunweiwei informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning AT liushengquan informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning AT liuyan informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning AT konglingqi informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning AT jianzhaorui informationextractionnetworkbasedonmultigranularityattentionandmultiscaleselflearning |