Cargando…
S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). How...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9030542/ https://www.ncbi.nlm.nih.gov/pubmed/35458837 http://dx.doi.org/10.3390/s22082852 |
_version_ | 1784692166560317440 |
---|---|
author | Yu, Jie Ji, Bin Li, Shasha Ma, Jun Liu, Huijun Xu, Hao |
author_facet | Yu, Jie Ji, Bin Li, Shasha Ma, Jun Liu, Huijun Xu, Hao |
author_sort | Yu, Jie |
collection | PubMed |
description | Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). However, these models are prone to cascade label misclassifications since a misclassified label results in incorrect label dependency, and so some following labels may also be misclassified. To address the above issue, we propose S-NER, a span-based NER model. To be specific, S-NER first splits raw texts into text spans and regards them as candidate entities; it then directly obtains the types of spans by conducting entity type classifications on span semantic representations, which eliminates the requirement for label dependency. Moreover, S-NER has a concise neural architecture in which it directly uses BERT as its encoder and a feed-forward network as its decoder. We evaluate S-NER on several benchmark datasets across three domains. Experimental results demonstrate that S-NER consistently outperforms the strongest baselines in terms of F1-score. Extensive analyses further confirm the efficacy of S-NER. |
format | Online Article Text |
id | pubmed-9030542 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-90305422022-04-23 S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition Yu, Jie Ji, Bin Li, Shasha Ma, Jun Liu, Huijun Xu, Hao Sensors (Basel) Communication Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). However, these models are prone to cascade label misclassifications since a misclassified label results in incorrect label dependency, and so some following labels may also be misclassified. To address the above issue, we propose S-NER, a span-based NER model. To be specific, S-NER first splits raw texts into text spans and regards them as candidate entities; it then directly obtains the types of spans by conducting entity type classifications on span semantic representations, which eliminates the requirement for label dependency. Moreover, S-NER has a concise neural architecture in which it directly uses BERT as its encoder and a feed-forward network as its decoder. We evaluate S-NER on several benchmark datasets across three domains. Experimental results demonstrate that S-NER consistently outperforms the strongest baselines in terms of F1-score. Extensive analyses further confirm the efficacy of S-NER. MDPI 2022-04-08 /pmc/articles/PMC9030542/ /pubmed/35458837 http://dx.doi.org/10.3390/s22082852 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Communication Yu, Jie Ji, Bin Li, Shasha Ma, Jun Liu, Huijun Xu, Hao S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition |
title | S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition |
title_full | S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition |
title_fullStr | S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition |
title_full_unstemmed | S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition |
title_short | S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition |
title_sort | s-ner: a concise and efficient span-based model for named entity recognition |
topic | Communication |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9030542/ https://www.ncbi.nlm.nih.gov/pubmed/35458837 http://dx.doi.org/10.3390/s22082852 |
work_keys_str_mv | AT yujie sneraconciseandefficientspanbasedmodelfornamedentityrecognition AT jibin sneraconciseandefficientspanbasedmodelfornamedentityrecognition AT lishasha sneraconciseandefficientspanbasedmodelfornamedentityrecognition AT majun sneraconciseandefficientspanbasedmodelfornamedentityrecognition AT liuhuijun sneraconciseandefficientspanbasedmodelfornamedentityrecognition AT xuhao sneraconciseandefficientspanbasedmodelfornamedentityrecognition |