Cargando…
Evaluating human versus machine learning performance in classifying research abstracts
We study whether humans or machine learning (ML) classification models are better at classifying scientific research abstracts according to a fixed set of discipline groups. We recruit both undergraduate and postgraduate assistants for this task in separate stages, and compare their performance agai...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7367789/ https://www.ncbi.nlm.nih.gov/pubmed/32836529 http://dx.doi.org/10.1007/s11192-020-03614-2 |
_version_ | 1783560483353133056 |
---|---|
author | Goh, Yeow Chong Cai, Xin Qing Theseira, Walter Ko, Giovanni Khor, Khiam Aik |
author_facet | Goh, Yeow Chong Cai, Xin Qing Theseira, Walter Ko, Giovanni Khor, Khiam Aik |
author_sort | Goh, Yeow Chong |
collection | PubMed |
description | We study whether humans or machine learning (ML) classification models are better at classifying scientific research abstracts according to a fixed set of discipline groups. We recruit both undergraduate and postgraduate assistants for this task in separate stages, and compare their performance against the support vectors machine ML algorithm at classifying European Research Council Starting Grant project abstracts to their actual evaluation panels, which are organised by discipline groups. On average, ML is more accurate than human classifiers, across a variety of training and test datasets, and across evaluation panels. ML classifiers trained on different training sets are also more reliable than human classifiers, meaning that different ML classifiers are more consistent in assigning the same classifications to any given abstract, compared to different human classifiers. While the top five percentile of human classifiers can outperform ML in limited cases, selection and training of such classifiers is likely costly and difficult compared to training ML models. Our results suggest ML models are a cost effective and highly accurate method for addressing problems in comparative bibliometric analysis, such as harmonising the discipline classifications of research from different funding agencies or countries. |
format | Online Article Text |
id | pubmed-7367789 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-73677892020-07-20 Evaluating human versus machine learning performance in classifying research abstracts Goh, Yeow Chong Cai, Xin Qing Theseira, Walter Ko, Giovanni Khor, Khiam Aik Scientometrics Article We study whether humans or machine learning (ML) classification models are better at classifying scientific research abstracts according to a fixed set of discipline groups. We recruit both undergraduate and postgraduate assistants for this task in separate stages, and compare their performance against the support vectors machine ML algorithm at classifying European Research Council Starting Grant project abstracts to their actual evaluation panels, which are organised by discipline groups. On average, ML is more accurate than human classifiers, across a variety of training and test datasets, and across evaluation panels. ML classifiers trained on different training sets are also more reliable than human classifiers, meaning that different ML classifiers are more consistent in assigning the same classifications to any given abstract, compared to different human classifiers. While the top five percentile of human classifiers can outperform ML in limited cases, selection and training of such classifiers is likely costly and difficult compared to training ML models. Our results suggest ML models are a cost effective and highly accurate method for addressing problems in comparative bibliometric analysis, such as harmonising the discipline classifications of research from different funding agencies or countries. Springer International Publishing 2020-07-18 2020 /pmc/articles/PMC7367789/ /pubmed/32836529 http://dx.doi.org/10.1007/s11192-020-03614-2 Text en © The Author(s) 2020 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Goh, Yeow Chong Cai, Xin Qing Theseira, Walter Ko, Giovanni Khor, Khiam Aik Evaluating human versus machine learning performance in classifying research abstracts |
title | Evaluating human versus machine learning performance in classifying research abstracts |
title_full | Evaluating human versus machine learning performance in classifying research abstracts |
title_fullStr | Evaluating human versus machine learning performance in classifying research abstracts |
title_full_unstemmed | Evaluating human versus machine learning performance in classifying research abstracts |
title_short | Evaluating human versus machine learning performance in classifying research abstracts |
title_sort | evaluating human versus machine learning performance in classifying research abstracts |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7367789/ https://www.ncbi.nlm.nih.gov/pubmed/32836529 http://dx.doi.org/10.1007/s11192-020-03614-2 |
work_keys_str_mv | AT gohyeowchong evaluatinghumanversusmachinelearningperformanceinclassifyingresearchabstracts AT caixinqing evaluatinghumanversusmachinelearningperformanceinclassifyingresearchabstracts AT theseirawalter evaluatinghumanversusmachinelearningperformanceinclassifyingresearchabstracts AT kogiovanni evaluatinghumanversusmachinelearningperformanceinclassifyingresearchabstracts AT khorkhiamaik evaluatinghumanversusmachinelearningperformanceinclassifyingresearchabstracts |