Cargando…
ANTIQUE: A Non-factoid Question Answering Benchmark
Considering the widespread use of mobile and voice search, answer passage retrieval for non-factoid questions plays a critical role in modern information retrieval systems. Despite the importance of the task, the community still feels the significant lack of large-scale non-factoid question answerin...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148024/ http://dx.doi.org/10.1007/978-3-030-45442-5_21 |
_version_ | 1783520514022572032 |
---|---|
author | Hashemi, Helia Aliannejadi, Mohammad Zamani, Hamed Croft, W. Bruce |
author_facet | Hashemi, Helia Aliannejadi, Mohammad Zamani, Hamed Croft, W. Bruce |
author_sort | Hashemi, Helia |
collection | PubMed |
description | Considering the widespread use of mobile and voice search, answer passage retrieval for non-factoid questions plays a critical role in modern information retrieval systems. Despite the importance of the task, the community still feels the significant lack of large-scale non-factoid question answering collections with real questions and comprehensive relevance judgments. In this paper, we develop and release a collection of 2,626 open-domain non-factoid questions from a diverse set of categories. The dataset, called ANTIQUE, contains 34k manual relevance annotations. The questions were asked by real users in a community question answering service, i.e., Yahoo! Answers. Relevance judgments for all the answers to each question were collected through crowdsourcing. To facilitate further research, we also include a brief analysis of the data as well as baseline results on both classical and neural IR models. |
format | Online Article Text |
id | pubmed-7148024 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-71480242020-04-13 ANTIQUE: A Non-factoid Question Answering Benchmark Hashemi, Helia Aliannejadi, Mohammad Zamani, Hamed Croft, W. Bruce Advances in Information Retrieval Article Considering the widespread use of mobile and voice search, answer passage retrieval for non-factoid questions plays a critical role in modern information retrieval systems. Despite the importance of the task, the community still feels the significant lack of large-scale non-factoid question answering collections with real questions and comprehensive relevance judgments. In this paper, we develop and release a collection of 2,626 open-domain non-factoid questions from a diverse set of categories. The dataset, called ANTIQUE, contains 34k manual relevance annotations. The questions were asked by real users in a community question answering service, i.e., Yahoo! Answers. Relevance judgments for all the answers to each question were collected through crowdsourcing. To facilitate further research, we also include a brief analysis of the data as well as baseline results on both classical and neural IR models. 2020-03-24 /pmc/articles/PMC7148024/ http://dx.doi.org/10.1007/978-3-030-45442-5_21 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Hashemi, Helia Aliannejadi, Mohammad Zamani, Hamed Croft, W. Bruce ANTIQUE: A Non-factoid Question Answering Benchmark |
title | ANTIQUE: A Non-factoid Question Answering Benchmark |
title_full | ANTIQUE: A Non-factoid Question Answering Benchmark |
title_fullStr | ANTIQUE: A Non-factoid Question Answering Benchmark |
title_full_unstemmed | ANTIQUE: A Non-factoid Question Answering Benchmark |
title_short | ANTIQUE: A Non-factoid Question Answering Benchmark |
title_sort | antique: a non-factoid question answering benchmark |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148024/ http://dx.doi.org/10.1007/978-3-030-45442-5_21 |
work_keys_str_mv | AT hashemihelia antiqueanonfactoidquestionansweringbenchmark AT aliannejadimohammad antiqueanonfactoidquestionansweringbenchmark AT zamanihamed antiqueanonfactoidquestionansweringbenchmark AT croftwbruce antiqueanonfactoidquestionansweringbenchmark |