Cargando…

BioVAE: a pre-trained latent variable language model for biomedical text mining

SUMMARY: Large-scale pre-trained language models (PLMs) have advanced state-of-the-art (SOTA) performance on various biomedical text mining tasks. The power of such PLMs can be combined with the advantages of deep generative models. These are examples of these combinations. However, they are trained...

Descripción completa

Detalles Bibliográficos
Autores principales: Trieu, Hai-Long, Miwa, Makoto, Ananiadou, Sophia
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8756089/
https://www.ncbi.nlm.nih.gov/pubmed/34636886
http://dx.doi.org/10.1093/bioinformatics/btab702
_version_ 1784632491054727168
author Trieu, Hai-Long
Miwa, Makoto
Ananiadou, Sophia
author_facet Trieu, Hai-Long
Miwa, Makoto
Ananiadou, Sophia
author_sort Trieu, Hai-Long
collection PubMed
description SUMMARY: Large-scale pre-trained language models (PLMs) have advanced state-of-the-art (SOTA) performance on various biomedical text mining tasks. The power of such PLMs can be combined with the advantages of deep generative models. These are examples of these combinations. However, they are trained only on general domain text, and biomedical models are still missing. In this work, we describe BioVAE, the first large-scale pre-trained latent variable language model for the biomedical domain, which uses the OPTIMUS framework to train on large volumes of biomedical text. The model shows SOTA performance on several biomedical text mining tasks when compared to existing publicly available biomedical PLMs. In addition, our model can generate more accurate biomedical sentences than the original OPTIMUS output. AVAILABILITY AND IMPLEMENTATION: Our source code and pre-trained models are freely available: https://github.com/aistairc/BioVAE. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
format Online
Article
Text
id pubmed-8756089
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-87560892022-01-13 BioVAE: a pre-trained latent variable language model for biomedical text mining Trieu, Hai-Long Miwa, Makoto Ananiadou, Sophia Bioinformatics Applications Notes SUMMARY: Large-scale pre-trained language models (PLMs) have advanced state-of-the-art (SOTA) performance on various biomedical text mining tasks. The power of such PLMs can be combined with the advantages of deep generative models. These are examples of these combinations. However, they are trained only on general domain text, and biomedical models are still missing. In this work, we describe BioVAE, the first large-scale pre-trained latent variable language model for the biomedical domain, which uses the OPTIMUS framework to train on large volumes of biomedical text. The model shows SOTA performance on several biomedical text mining tasks when compared to existing publicly available biomedical PLMs. In addition, our model can generate more accurate biomedical sentences than the original OPTIMUS output. AVAILABILITY AND IMPLEMENTATION: Our source code and pre-trained models are freely available: https://github.com/aistairc/BioVAE. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online. Oxford University Press 2021-10-12 /pmc/articles/PMC8756089/ /pubmed/34636886 http://dx.doi.org/10.1093/bioinformatics/btab702 Text en © The Author(s) 2021. Published by Oxford University Press. https://creativecommons.org/licenses/by-nc/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com
spellingShingle Applications Notes
Trieu, Hai-Long
Miwa, Makoto
Ananiadou, Sophia
BioVAE: a pre-trained latent variable language model for biomedical text mining
title BioVAE: a pre-trained latent variable language model for biomedical text mining
title_full BioVAE: a pre-trained latent variable language model for biomedical text mining
title_fullStr BioVAE: a pre-trained latent variable language model for biomedical text mining
title_full_unstemmed BioVAE: a pre-trained latent variable language model for biomedical text mining
title_short BioVAE: a pre-trained latent variable language model for biomedical text mining
title_sort biovae: a pre-trained latent variable language model for biomedical text mining
topic Applications Notes
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8756089/
https://www.ncbi.nlm.nih.gov/pubmed/34636886
http://dx.doi.org/10.1093/bioinformatics/btab702
work_keys_str_mv AT trieuhailong biovaeapretrainedlatentvariablelanguagemodelforbiomedicaltextmining
AT miwamakoto biovaeapretrainedlatentvariablelanguagemodelforbiomedicaltextmining
AT ananiadousophia biovaeapretrainedlatentvariablelanguagemodelforbiomedicaltextmining