Cargando…

S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition

Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). How...

Descripción completa

Detalles Bibliográficos
Autores principales: Yu, Jie, Ji, Bin, Li, Shasha, Ma, Jun, Liu, Huijun, Xu, Hao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9030542/
https://www.ncbi.nlm.nih.gov/pubmed/35458837
http://dx.doi.org/10.3390/s22082852
_version_ 1784692166560317440
author Yu, Jie
Ji, Bin
Li, Shasha
Ma, Jun
Liu, Huijun
Xu, Hao
author_facet Yu, Jie
Ji, Bin
Li, Shasha
Ma, Jun
Liu, Huijun
Xu, Hao
author_sort Yu, Jie
collection PubMed
description Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). However, these models are prone to cascade label misclassifications since a misclassified label results in incorrect label dependency, and so some following labels may also be misclassified. To address the above issue, we propose S-NER, a span-based NER model. To be specific, S-NER first splits raw texts into text spans and regards them as candidate entities; it then directly obtains the types of spans by conducting entity type classifications on span semantic representations, which eliminates the requirement for label dependency. Moreover, S-NER has a concise neural architecture in which it directly uses BERT as its encoder and a feed-forward network as its decoder. We evaluate S-NER on several benchmark datasets across three domains. Experimental results demonstrate that S-NER consistently outperforms the strongest baselines in terms of F1-score. Extensive analyses further confirm the efficacy of S-NER.
format Online
Article
Text
id pubmed-9030542
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-90305422022-04-23 S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition Yu, Jie Ji, Bin Li, Shasha Ma, Jun Liu, Huijun Xu, Hao Sensors (Basel) Communication Named entity recognition (NER) is a task that seeks to recognize entities in raw texts and is a precondition for a series of downstream NLP tasks. Traditionally, prior NER models use the sequence labeling mechanism which requires label dependency captured by the conditional random fields (CRFs). However, these models are prone to cascade label misclassifications since a misclassified label results in incorrect label dependency, and so some following labels may also be misclassified. To address the above issue, we propose S-NER, a span-based NER model. To be specific, S-NER first splits raw texts into text spans and regards them as candidate entities; it then directly obtains the types of spans by conducting entity type classifications on span semantic representations, which eliminates the requirement for label dependency. Moreover, S-NER has a concise neural architecture in which it directly uses BERT as its encoder and a feed-forward network as its decoder. We evaluate S-NER on several benchmark datasets across three domains. Experimental results demonstrate that S-NER consistently outperforms the strongest baselines in terms of F1-score. Extensive analyses further confirm the efficacy of S-NER. MDPI 2022-04-08 /pmc/articles/PMC9030542/ /pubmed/35458837 http://dx.doi.org/10.3390/s22082852 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Communication
Yu, Jie
Ji, Bin
Li, Shasha
Ma, Jun
Liu, Huijun
Xu, Hao
S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
title S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
title_full S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
title_fullStr S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
title_full_unstemmed S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
title_short S-NER: A Concise and Efficient Span-Based Model for Named Entity Recognition
title_sort s-ner: a concise and efficient span-based model for named entity recognition
topic Communication
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9030542/
https://www.ncbi.nlm.nih.gov/pubmed/35458837
http://dx.doi.org/10.3390/s22082852
work_keys_str_mv AT yujie sneraconciseandefficientspanbasedmodelfornamedentityrecognition
AT jibin sneraconciseandefficientspanbasedmodelfornamedentityrecognition
AT lishasha sneraconciseandefficientspanbasedmodelfornamedentityrecognition
AT majun sneraconciseandefficientspanbasedmodelfornamedentityrecognition
AT liuhuijun sneraconciseandefficientspanbasedmodelfornamedentityrecognition
AT xuhao sneraconciseandefficientspanbasedmodelfornamedentityrecognition