Cargando…
Bioformer: an efficient transformer language model for biomedical text mining
Pretrained language models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance in natural language processing (NLP) tasks. Recently, BERT has been adapted to the biomedical domain. Despite the effectiveness, these models have hundreds of...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Cornell University
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10029052/ https://www.ncbi.nlm.nih.gov/pubmed/36945685 |
_version_ | 1784910066988613632 |
---|---|
author | Fang, Li Chen, Qingyu Wei, Chih-Hsuan Lu, Zhiyong Wang, Kai |
author_facet | Fang, Li Chen, Qingyu Wei, Chih-Hsuan Lu, Zhiyong Wang, Kai |
author_sort | Fang, Li |
collection | PubMed |
description | Pretrained language models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance in natural language processing (NLP) tasks. Recently, BERT has been adapted to the biomedical domain. Despite the effectiveness, these models have hundreds of millions of parameters and are computationally expensive when applied to large-scale NLP applications. We hypothesized that the number of parameters of the original BERT can be dramatically reduced with minor impact on performance. In this study, we present Bioformer, a compact BERT model for biomedical text mining. We pretrained two Bioformer models (named Bioformer(8L) and Bioformer(16L)) which reduced the model size by 60% compared to BERT(Base). Bioformer uses a biomedical vocabulary and was pre-trained from scratch on PubMed abstracts and PubMed Central full-text articles. We thoroughly evaluated the performance of Bioformer as well as existing biomedical BERT models including BioBERT and PubMedBERT on 15 benchmark datasets of four different biomedical NLP tasks: named entity recognition, relation extraction, question answering and document classification. The results show that with 60% fewer parameters, Bioformer(16L) is only 0.1% less accurate than PubMedBERT while Bioformer(8L) is 0.9% less accurate than PubMedBERT. Both Bioformer(16L) and Bioformer(8L) outperformed BioBERT(Base-v1.1). In addition, Bioformer(16L) and Bioformer(8L) are 2–3 fold as fast as PubMedBERT/BioBERT(Base-v1.1). Bioformer has been successfully deployed to PubTator Central providing gene annotations over 35 million PubMed abstracts and 5 million PubMed Central full-text articles. We make Bioformer publicly available via https://github.com/WGLab/bioformer, including pre-trained models, datasets, and instructions for downstream use. |
format | Online Article Text |
id | pubmed-10029052 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Cornell University |
record_format | MEDLINE/PubMed |
spelling | pubmed-100290522023-03-22 Bioformer: an efficient transformer language model for biomedical text mining Fang, Li Chen, Qingyu Wei, Chih-Hsuan Lu, Zhiyong Wang, Kai ArXiv Article Pretrained language models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance in natural language processing (NLP) tasks. Recently, BERT has been adapted to the biomedical domain. Despite the effectiveness, these models have hundreds of millions of parameters and are computationally expensive when applied to large-scale NLP applications. We hypothesized that the number of parameters of the original BERT can be dramatically reduced with minor impact on performance. In this study, we present Bioformer, a compact BERT model for biomedical text mining. We pretrained two Bioformer models (named Bioformer(8L) and Bioformer(16L)) which reduced the model size by 60% compared to BERT(Base). Bioformer uses a biomedical vocabulary and was pre-trained from scratch on PubMed abstracts and PubMed Central full-text articles. We thoroughly evaluated the performance of Bioformer as well as existing biomedical BERT models including BioBERT and PubMedBERT on 15 benchmark datasets of four different biomedical NLP tasks: named entity recognition, relation extraction, question answering and document classification. The results show that with 60% fewer parameters, Bioformer(16L) is only 0.1% less accurate than PubMedBERT while Bioformer(8L) is 0.9% less accurate than PubMedBERT. Both Bioformer(16L) and Bioformer(8L) outperformed BioBERT(Base-v1.1). In addition, Bioformer(16L) and Bioformer(8L) are 2–3 fold as fast as PubMedBERT/BioBERT(Base-v1.1). Bioformer has been successfully deployed to PubTator Central providing gene annotations over 35 million PubMed abstracts and 5 million PubMed Central full-text articles. We make Bioformer publicly available via https://github.com/WGLab/bioformer, including pre-trained models, datasets, and instructions for downstream use. Cornell University 2023-02-03 /pmc/articles/PMC10029052/ /pubmed/36945685 Text en https://creativecommons.org/licenses/by/4.0/This work is licensed under a Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which allows reusers to distribute, remix, adapt, and build upon the material in any medium or format, so long as attribution is given to the creator. The license allows for commercial use. |
spellingShingle | Article Fang, Li Chen, Qingyu Wei, Chih-Hsuan Lu, Zhiyong Wang, Kai Bioformer: an efficient transformer language model for biomedical text mining |
title | Bioformer: an efficient transformer language model for biomedical text mining |
title_full | Bioformer: an efficient transformer language model for biomedical text mining |
title_fullStr | Bioformer: an efficient transformer language model for biomedical text mining |
title_full_unstemmed | Bioformer: an efficient transformer language model for biomedical text mining |
title_short | Bioformer: an efficient transformer language model for biomedical text mining |
title_sort | bioformer: an efficient transformer language model for biomedical text mining |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10029052/ https://www.ncbi.nlm.nih.gov/pubmed/36945685 |
work_keys_str_mv | AT fangli bioformeranefficienttransformerlanguagemodelforbiomedicaltextmining AT chenqingyu bioformeranefficienttransformerlanguagemodelforbiomedicaltextmining AT weichihhsuan bioformeranefficienttransformerlanguagemodelforbiomedicaltextmining AT luzhiyong bioformeranefficienttransformerlanguagemodelforbiomedicaltextmining AT wangkai bioformeranefficienttransformerlanguagemodelforbiomedicaltextmining |