Cargando…

LitCovid ensemble learning for COVID-19 multi-label classification

The Coronavirus Disease 2019 (COVID-19) pandemic has shifted the focus of research worldwide, and more than 10 000 new articles per month have concentrated on COVID-19–related topics. Considering this rapidly growing literature, the efficient and precise extraction of the main topics of COVID-19–rel...

Descripción completa

Detalles Bibliográficos
Autores principales: Gu, Jinghang, Chersoni, Emmanuele, Wang, Xing, Huang, Chu-Ren, Qian, Longhua, Zhou, Guodong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9693804/
https://www.ncbi.nlm.nih.gov/pubmed/36426767
http://dx.doi.org/10.1093/database/baac103
_version_ 1784837636567859200
author Gu, Jinghang
Chersoni, Emmanuele
Wang, Xing
Huang, Chu-Ren
Qian, Longhua
Zhou, Guodong
author_facet Gu, Jinghang
Chersoni, Emmanuele
Wang, Xing
Huang, Chu-Ren
Qian, Longhua
Zhou, Guodong
author_sort Gu, Jinghang
collection PubMed
description The Coronavirus Disease 2019 (COVID-19) pandemic has shifted the focus of research worldwide, and more than 10 000 new articles per month have concentrated on COVID-19–related topics. Considering this rapidly growing literature, the efficient and precise extraction of the main topics of COVID-19–relevant articles is of great importance. The manual curation of this information for biomedical literature is labor-intensive and time-consuming, and as such the procedure is insufficient and difficult to maintain. In response to these complications, the BioCreative VII community has proposed a challenging task, LitCovid Track, calling for a global effort to automatically extract semantic topics for COVID-19 literature. This article describes our work on the BioCreative VII LitCovid Track. We proposed the LitCovid Ensemble Learning (LCEL) method for the tasks and integrated multiple biomedical pretrained models to address the COVID-19 multi-label classification problem. Specifically, seven different transformer-based pretrained models were ensembled for the initialization and fine-tuning processes independently. To enhance the representation abilities of the deep neural models, diverse additional biomedical knowledge was utilized to facilitate the fruitfulness of the semantic expressions. Simple yet effective data augmentation was also leveraged to address the learning deficiency during the training phase. In addition, given the imbalanced label distribution of the challenging task, a novel asymmetric loss function was applied to the LCEL model, which explicitly adjusted the negative–positive importance by assigning different exponential decay factors and helped the model focus on the positive samples. After the training phase, an ensemble bagging strategy was adopted to merge the outputs from each model for final predictions. The experimental results show the effectiveness of our proposed approach, as LCEL obtains the state-of-the-art performance on the LitCovid dataset. Database URL: https://github.com/JHnlp/LCEL
format Online
Article
Text
id pubmed-9693804
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-96938042022-11-28 LitCovid ensemble learning for COVID-19 multi-label classification Gu, Jinghang Chersoni, Emmanuele Wang, Xing Huang, Chu-Ren Qian, Longhua Zhou, Guodong Database (Oxford) Original Article The Coronavirus Disease 2019 (COVID-19) pandemic has shifted the focus of research worldwide, and more than 10 000 new articles per month have concentrated on COVID-19–related topics. Considering this rapidly growing literature, the efficient and precise extraction of the main topics of COVID-19–relevant articles is of great importance. The manual curation of this information for biomedical literature is labor-intensive and time-consuming, and as such the procedure is insufficient and difficult to maintain. In response to these complications, the BioCreative VII community has proposed a challenging task, LitCovid Track, calling for a global effort to automatically extract semantic topics for COVID-19 literature. This article describes our work on the BioCreative VII LitCovid Track. We proposed the LitCovid Ensemble Learning (LCEL) method for the tasks and integrated multiple biomedical pretrained models to address the COVID-19 multi-label classification problem. Specifically, seven different transformer-based pretrained models were ensembled for the initialization and fine-tuning processes independently. To enhance the representation abilities of the deep neural models, diverse additional biomedical knowledge was utilized to facilitate the fruitfulness of the semantic expressions. Simple yet effective data augmentation was also leveraged to address the learning deficiency during the training phase. In addition, given the imbalanced label distribution of the challenging task, a novel asymmetric loss function was applied to the LCEL model, which explicitly adjusted the negative–positive importance by assigning different exponential decay factors and helped the model focus on the positive samples. After the training phase, an ensemble bagging strategy was adopted to merge the outputs from each model for final predictions. The experimental results show the effectiveness of our proposed approach, as LCEL obtains the state-of-the-art performance on the LitCovid dataset. Database URL: https://github.com/JHnlp/LCEL Oxford University Press 2022-11-25 /pmc/articles/PMC9693804/ /pubmed/36426767 http://dx.doi.org/10.1093/database/baac103 Text en © The Author(s) 2022. Published by Oxford University Press. https://creativecommons.org/licenses/by-nc/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com
spellingShingle Original Article
Gu, Jinghang
Chersoni, Emmanuele
Wang, Xing
Huang, Chu-Ren
Qian, Longhua
Zhou, Guodong
LitCovid ensemble learning for COVID-19 multi-label classification
title LitCovid ensemble learning for COVID-19 multi-label classification
title_full LitCovid ensemble learning for COVID-19 multi-label classification
title_fullStr LitCovid ensemble learning for COVID-19 multi-label classification
title_full_unstemmed LitCovid ensemble learning for COVID-19 multi-label classification
title_short LitCovid ensemble learning for COVID-19 multi-label classification
title_sort litcovid ensemble learning for covid-19 multi-label classification
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9693804/
https://www.ncbi.nlm.nih.gov/pubmed/36426767
http://dx.doi.org/10.1093/database/baac103
work_keys_str_mv AT gujinghang litcovidensemblelearningforcovid19multilabelclassification
AT chersoniemmanuele litcovidensemblelearningforcovid19multilabelclassification
AT wangxing litcovidensemblelearningforcovid19multilabelclassification
AT huangchuren litcovidensemblelearningforcovid19multilabelclassification
AT qianlonghua litcovidensemblelearningforcovid19multilabelclassification
AT zhouguodong litcovidensemblelearningforcovid19multilabelclassification