Cargando…
LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation
The rapid growth of biomedical literature poses a significant challenge for curation and interpretation. This has become more evident during the COVID-19 pandemic. LitCovid, a literature database of COVID-19 related papers in PubMed, has accumulated over 200,000 articles with millions of accesses. A...
Formato: | Online Artículo Texto |
---|---|
Lenguaje: | English |
Publicado: |
IEEE
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9647722/ https://www.ncbi.nlm.nih.gov/pubmed/35536809 http://dx.doi.org/10.1109/TCBB.2022.3173562 |
_version_ | 1784827437320765440 |
---|---|
collection | PubMed |
description | The rapid growth of biomedical literature poses a significant challenge for curation and interpretation. This has become more evident during the COVID-19 pandemic. LitCovid, a literature database of COVID-19 related papers in PubMed, has accumulated over 200,000 articles with millions of accesses. Approximately 10,000 new articles are added to LitCovid every month. A main curation task in LitCovid is topic annotation where an article is assigned with up to eight topics, e.g., Treatment and Diagnosis. The annotated topics have been widely used both in LitCovid (e.g., accounting for ∼18% of total uses) and downstream studies such as network generation. However, it has been a primary curation bottleneck due to the nature of the task and the rapid literature growth. This study proposes LITMC-BERT, a transformer-based multi-label classification method in biomedical literature. It uses a shared transformer backbone for all the labels while also captures label-specific features and the correlations between label pairs. We compare LITMC-BERT with three baseline models on two datasets. Its micro-F1 and instance-based F1 are 5% and 4% higher than the current best results, respectively, and only requires ∼18% of the inference time than the Binary BERT baseline. The related datasets and models are available via https://github.com/ncbi/ml-transformer. |
format | Online Article Text |
id | pubmed-9647722 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | IEEE |
record_format | MEDLINE/PubMed |
spelling | pubmed-96477222022-11-14 LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation IEEE/ACM Trans Comput Biol Bioinform Article The rapid growth of biomedical literature poses a significant challenge for curation and interpretation. This has become more evident during the COVID-19 pandemic. LitCovid, a literature database of COVID-19 related papers in PubMed, has accumulated over 200,000 articles with millions of accesses. Approximately 10,000 new articles are added to LitCovid every month. A main curation task in LitCovid is topic annotation where an article is assigned with up to eight topics, e.g., Treatment and Diagnosis. The annotated topics have been widely used both in LitCovid (e.g., accounting for ∼18% of total uses) and downstream studies such as network generation. However, it has been a primary curation bottleneck due to the nature of the task and the rapid literature growth. This study proposes LITMC-BERT, a transformer-based multi-label classification method in biomedical literature. It uses a shared transformer backbone for all the labels while also captures label-specific features and the correlations between label pairs. We compare LITMC-BERT with three baseline models on two datasets. Its micro-F1 and instance-based F1 are 5% and 4% higher than the current best results, respectively, and only requires ∼18% of the inference time than the Binary BERT baseline. The related datasets and models are available via https://github.com/ncbi/ml-transformer. IEEE 2022-05-10 /pmc/articles/PMC9647722/ /pubmed/35536809 http://dx.doi.org/10.1109/TCBB.2022.3173562 Text en https://creativecommons.org/licenses/by/4.0/This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/ |
spellingShingle | Article LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation |
title | LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation |
title_full | LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation |
title_fullStr | LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation |
title_full_unstemmed | LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation |
title_short | LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation |
title_sort | litmc-bert: transformer-based multi-label classification of biomedical literature with an application on covid-19 literature curation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9647722/ https://www.ncbi.nlm.nih.gov/pubmed/35536809 http://dx.doi.org/10.1109/TCBB.2022.3173562 |
work_keys_str_mv | AT litmcberttransformerbasedmultilabelclassificationofbiomedicalliteraturewithanapplicationoncovid19literaturecuration AT litmcberttransformerbasedmultilabelclassificationofbiomedicalliteraturewithanapplicationoncovid19literaturecuration AT litmcberttransformerbasedmultilabelclassificationofbiomedicalliteraturewithanapplicationoncovid19literaturecuration AT litmcberttransformerbasedmultilabelclassificationofbiomedicalliteraturewithanapplicationoncovid19literaturecuration |