Cargando…
ExpMRC: explainability evaluation for machine reading comprehension
Achieving human-level performance on some Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, it is necessary to provide both answer prediction and its explanation to further improve the MRC system's reliab...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9048090/ https://www.ncbi.nlm.nih.gov/pubmed/35497046 http://dx.doi.org/10.1016/j.heliyon.2022.e09290 |
_version_ | 1784695866396770304 |
---|---|
author | Cui, Yiming Liu, Ting Che, Wanxiang Chen, Zhigang Wang, Shijin |
author_facet | Cui, Yiming Liu, Ting Che, Wanxiang Chen, Zhigang Wang, Shijin |
author_sort | Cui, Yiming |
collection | PubMed |
description | Achieving human-level performance on some Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, it is necessary to provide both answer prediction and its explanation to further improve the MRC system's reliability, especially for real-life applications. In this paper, we propose a new benchmark called ExpMRC for evaluating the textual explainability of the MRC systems. ExpMRC contains four subsets, including SQuAD, CMRC 2018, RACE(+), and C(3), with additional annotations of the answer's evidence. The MRC systems are required to give not only the correct answer but also its explanation. We use state-of-the-art PLMs to build baseline systems and adopt various unsupervised approaches to extract both answer and evidence spans without human-annotated evidence spans. The experimental results show that these models are still far from human performance, suggesting that the ExpMRC is challenging. Resources (data and baselines) are available through https://github.com/ymcui/expmrc. |
format | Online Article Text |
id | pubmed-9048090 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-90480902022-04-29 ExpMRC: explainability evaluation for machine reading comprehension Cui, Yiming Liu, Ting Che, Wanxiang Chen, Zhigang Wang, Shijin Heliyon Research Article Achieving human-level performance on some Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, it is necessary to provide both answer prediction and its explanation to further improve the MRC system's reliability, especially for real-life applications. In this paper, we propose a new benchmark called ExpMRC for evaluating the textual explainability of the MRC systems. ExpMRC contains four subsets, including SQuAD, CMRC 2018, RACE(+), and C(3), with additional annotations of the answer's evidence. The MRC systems are required to give not only the correct answer but also its explanation. We use state-of-the-art PLMs to build baseline systems and adopt various unsupervised approaches to extract both answer and evidence spans without human-annotated evidence spans. The experimental results show that these models are still far from human performance, suggesting that the ExpMRC is challenging. Resources (data and baselines) are available through https://github.com/ymcui/expmrc. Elsevier 2022-04-19 /pmc/articles/PMC9048090/ /pubmed/35497046 http://dx.doi.org/10.1016/j.heliyon.2022.e09290 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Research Article Cui, Yiming Liu, Ting Che, Wanxiang Chen, Zhigang Wang, Shijin ExpMRC: explainability evaluation for machine reading comprehension |
title | ExpMRC: explainability evaluation for machine reading comprehension |
title_full | ExpMRC: explainability evaluation for machine reading comprehension |
title_fullStr | ExpMRC: explainability evaluation for machine reading comprehension |
title_full_unstemmed | ExpMRC: explainability evaluation for machine reading comprehension |
title_short | ExpMRC: explainability evaluation for machine reading comprehension |
title_sort | expmrc: explainability evaluation for machine reading comprehension |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9048090/ https://www.ncbi.nlm.nih.gov/pubmed/35497046 http://dx.doi.org/10.1016/j.heliyon.2022.e09290 |
work_keys_str_mv | AT cuiyiming expmrcexplainabilityevaluationformachinereadingcomprehension AT liuting expmrcexplainabilityevaluationformachinereadingcomprehension AT chewanxiang expmrcexplainabilityevaluationformachinereadingcomprehension AT chenzhigang expmrcexplainabilityevaluationformachinereadingcomprehension AT wangshijin expmrcexplainabilityevaluationformachinereadingcomprehension |