Cargando…

ExpMRC: explainability evaluation for machine reading comprehension

Achieving human-level performance on some Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, it is necessary to provide both answer prediction and its explanation to further improve the MRC system's reliab...

Descripción completa

Detalles Bibliográficos
Autores principales: Cui, Yiming, Liu, Ting, Che, Wanxiang, Chen, Zhigang, Wang, Shijin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9048090/
https://www.ncbi.nlm.nih.gov/pubmed/35497046
http://dx.doi.org/10.1016/j.heliyon.2022.e09290
_version_ 1784695866396770304
author Cui, Yiming
Liu, Ting
Che, Wanxiang
Chen, Zhigang
Wang, Shijin
author_facet Cui, Yiming
Liu, Ting
Che, Wanxiang
Chen, Zhigang
Wang, Shijin
author_sort Cui, Yiming
collection PubMed
description Achieving human-level performance on some Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, it is necessary to provide both answer prediction and its explanation to further improve the MRC system's reliability, especially for real-life applications. In this paper, we propose a new benchmark called ExpMRC for evaluating the textual explainability of the MRC systems. ExpMRC contains four subsets, including SQuAD, CMRC 2018, RACE(+), and C(3), with additional annotations of the answer's evidence. The MRC systems are required to give not only the correct answer but also its explanation. We use state-of-the-art PLMs to build baseline systems and adopt various unsupervised approaches to extract both answer and evidence spans without human-annotated evidence spans. The experimental results show that these models are still far from human performance, suggesting that the ExpMRC is challenging. Resources (data and baselines) are available through https://github.com/ymcui/expmrc.
format Online
Article
Text
id pubmed-9048090
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-90480902022-04-29 ExpMRC: explainability evaluation for machine reading comprehension Cui, Yiming Liu, Ting Che, Wanxiang Chen, Zhigang Wang, Shijin Heliyon Research Article Achieving human-level performance on some Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, it is necessary to provide both answer prediction and its explanation to further improve the MRC system's reliability, especially for real-life applications. In this paper, we propose a new benchmark called ExpMRC for evaluating the textual explainability of the MRC systems. ExpMRC contains four subsets, including SQuAD, CMRC 2018, RACE(+), and C(3), with additional annotations of the answer's evidence. The MRC systems are required to give not only the correct answer but also its explanation. We use state-of-the-art PLMs to build baseline systems and adopt various unsupervised approaches to extract both answer and evidence spans without human-annotated evidence spans. The experimental results show that these models are still far from human performance, suggesting that the ExpMRC is challenging. Resources (data and baselines) are available through https://github.com/ymcui/expmrc. Elsevier 2022-04-19 /pmc/articles/PMC9048090/ /pubmed/35497046 http://dx.doi.org/10.1016/j.heliyon.2022.e09290 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Research Article
Cui, Yiming
Liu, Ting
Che, Wanxiang
Chen, Zhigang
Wang, Shijin
ExpMRC: explainability evaluation for machine reading comprehension
title ExpMRC: explainability evaluation for machine reading comprehension
title_full ExpMRC: explainability evaluation for machine reading comprehension
title_fullStr ExpMRC: explainability evaluation for machine reading comprehension
title_full_unstemmed ExpMRC: explainability evaluation for machine reading comprehension
title_short ExpMRC: explainability evaluation for machine reading comprehension
title_sort expmrc: explainability evaluation for machine reading comprehension
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9048090/
https://www.ncbi.nlm.nih.gov/pubmed/35497046
http://dx.doi.org/10.1016/j.heliyon.2022.e09290
work_keys_str_mv AT cuiyiming expmrcexplainabilityevaluationformachinereadingcomprehension
AT liuting expmrcexplainabilityevaluationformachinereadingcomprehension
AT chewanxiang expmrcexplainabilityevaluationformachinereadingcomprehension
AT chenzhigang expmrcexplainabilityevaluationformachinereadingcomprehension
AT wangshijin expmrcexplainabilityevaluationformachinereadingcomprehension