Cargando…
Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
In this paper, we focus on the problem of question retrieval in community Question Answering (cQA) which aims to retrieve from the community archives the previous questions that are semantically equivalent to the new queries. The major challenges in this crucial task are the shortness of the questio...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7298191/ http://dx.doi.org/10.1007/978-3-030-51310-8_23 |
_version_ | 1783547166200954880 |
---|---|
author | Othman, Nouha Faiz, Rim Smaïli, Kamel |
author_facet | Othman, Nouha Faiz, Rim Smaïli, Kamel |
author_sort | Othman, Nouha |
collection | PubMed |
description | In this paper, we focus on the problem of question retrieval in community Question Answering (cQA) which aims to retrieve from the community archives the previous questions that are semantically equivalent to the new queries. The major challenges in this crucial task are the shortness of the questions as well as the word mismatch problem as users can formulate the same query using different wording. While numerous attempts have been made to address this problem, most existing methods relied on supervised models which significantly depend on large training data sets and manual feature engineering. Such methods are mostly constrained by their specificities that put aside the word order and ignore syntactic and semantic relationships. In this work, we rely on Neural Networks (NNs) which can learn rich dense representations of text data and enable the prediction of the textual similarity between the community questions. We propose a deep learning approach based on a Siamese architecture with LSTM networks, augmented with an attention mechanism. We test different similarity measures to predict the semantic similarity between the community questions. Experiments conducted on real cQA data sets in English and Arabic show that the performance of question retrieval is improved as compared to other competitive methods. |
format | Online Article Text |
id | pubmed-7298191 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72981912020-06-17 Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM Othman, Nouha Faiz, Rim Smaïli, Kamel Natural Language Processing and Information Systems Article In this paper, we focus on the problem of question retrieval in community Question Answering (cQA) which aims to retrieve from the community archives the previous questions that are semantically equivalent to the new queries. The major challenges in this crucial task are the shortness of the questions as well as the word mismatch problem as users can formulate the same query using different wording. While numerous attempts have been made to address this problem, most existing methods relied on supervised models which significantly depend on large training data sets and manual feature engineering. Such methods are mostly constrained by their specificities that put aside the word order and ignore syntactic and semantic relationships. In this work, we rely on Neural Networks (NNs) which can learn rich dense representations of text data and enable the prediction of the textual similarity between the community questions. We propose a deep learning approach based on a Siamese architecture with LSTM networks, augmented with an attention mechanism. We test different similarity measures to predict the semantic similarity between the community questions. Experiments conducted on real cQA data sets in English and Arabic show that the performance of question retrieval is improved as compared to other competitive methods. 2020-05-26 /pmc/articles/PMC7298191/ http://dx.doi.org/10.1007/978-3-030-51310-8_23 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Othman, Nouha Faiz, Rim Smaïli, Kamel Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM |
title | Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM |
title_full | Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM |
title_fullStr | Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM |
title_full_unstemmed | Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM |
title_short | Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM |
title_sort | improving the community question retrieval performance using attention-based siamese lstm |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7298191/ http://dx.doi.org/10.1007/978-3-030-51310-8_23 |
work_keys_str_mv | AT othmannouha improvingthecommunityquestionretrievalperformanceusingattentionbasedsiameselstm AT faizrim improvingthecommunityquestionretrievalperformanceusingattentionbasedsiameselstm AT smailikamel improvingthecommunityquestionretrievalperformanceusingattentionbasedsiameselstm |