Cargando…
Investigating Query Expansion and Coreference Resolution in Question Answering on BERT
The Bidirectional Encoder Representations from Transformers (BERT) model produces state-of-the-art results in many question answering (QA) datasets, including the Stanford Question Answering Dataset (SQuAD). This paper presents a query expansion (QE) method that identifies good terms from input ques...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7298170/ http://dx.doi.org/10.1007/978-3-030-51310-8_5 |
_version_ | 1783547161229656064 |
---|---|
author | Bhattacharjee, Santanu Haque, Rejwanul de Buy Wenniger, Gideon Maillette Way, Andy |
author_facet | Bhattacharjee, Santanu Haque, Rejwanul de Buy Wenniger, Gideon Maillette Way, Andy |
author_sort | Bhattacharjee, Santanu |
collection | PubMed |
description | The Bidirectional Encoder Representations from Transformers (BERT) model produces state-of-the-art results in many question answering (QA) datasets, including the Stanford Question Answering Dataset (SQuAD). This paper presents a query expansion (QE) method that identifies good terms from input questions, extracts synonyms for the good terms using a widely-used language resource, WordNet, and selects the most relevant synonyms from the list of extracted synonyms. The paper also introduces a novel QE method that produces many alternative sequences for a given input question using same-language machine translation (MT). Furthermore, we use a coreference resolution (CR) technique to identify anaphors or cataphors in paragraphs and substitute them with the original referents. We found that the QA system with this simple CR technique significantly outperforms the BERT baseline in a QA task. We also found that our best-performing QA system is the one that applies these three preprocessing methods (two QE and CR methods) together to BERT, which produces an excellent [Formula: see text] score (89.8 [Formula: see text] points) in a QA task. Further, we present a comparative analysis on the performances of the BERT QA models taking a variety of criteria into account, and demonstrate our findings in the answer span prediction task. |
format | Online Article Text |
id | pubmed-7298170 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72981702020-06-17 Investigating Query Expansion and Coreference Resolution in Question Answering on BERT Bhattacharjee, Santanu Haque, Rejwanul de Buy Wenniger, Gideon Maillette Way, Andy Natural Language Processing and Information Systems Article The Bidirectional Encoder Representations from Transformers (BERT) model produces state-of-the-art results in many question answering (QA) datasets, including the Stanford Question Answering Dataset (SQuAD). This paper presents a query expansion (QE) method that identifies good terms from input questions, extracts synonyms for the good terms using a widely-used language resource, WordNet, and selects the most relevant synonyms from the list of extracted synonyms. The paper also introduces a novel QE method that produces many alternative sequences for a given input question using same-language machine translation (MT). Furthermore, we use a coreference resolution (CR) technique to identify anaphors or cataphors in paragraphs and substitute them with the original referents. We found that the QA system with this simple CR technique significantly outperforms the BERT baseline in a QA task. We also found that our best-performing QA system is the one that applies these three preprocessing methods (two QE and CR methods) together to BERT, which produces an excellent [Formula: see text] score (89.8 [Formula: see text] points) in a QA task. Further, we present a comparative analysis on the performances of the BERT QA models taking a variety of criteria into account, and demonstrate our findings in the answer span prediction task. 2020-05-26 /pmc/articles/PMC7298170/ http://dx.doi.org/10.1007/978-3-030-51310-8_5 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Bhattacharjee, Santanu Haque, Rejwanul de Buy Wenniger, Gideon Maillette Way, Andy Investigating Query Expansion and Coreference Resolution in Question Answering on BERT |
title | Investigating Query Expansion and Coreference Resolution in Question Answering on BERT |
title_full | Investigating Query Expansion and Coreference Resolution in Question Answering on BERT |
title_fullStr | Investigating Query Expansion and Coreference Resolution in Question Answering on BERT |
title_full_unstemmed | Investigating Query Expansion and Coreference Resolution in Question Answering on BERT |
title_short | Investigating Query Expansion and Coreference Resolution in Question Answering on BERT |
title_sort | investigating query expansion and coreference resolution in question answering on bert |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7298170/ http://dx.doi.org/10.1007/978-3-030-51310-8_5 |
work_keys_str_mv | AT bhattacharjeesantanu investigatingqueryexpansionandcoreferenceresolutioninquestionansweringonbert AT haquerejwanul investigatingqueryexpansionandcoreferenceresolutioninquestionansweringonbert AT debuywennigergideonmaillette investigatingqueryexpansionandcoreferenceresolutioninquestionansweringonbert AT wayandy investigatingqueryexpansionandcoreferenceresolutioninquestionansweringonbert |