Cargando…

Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM

In this paper, we focus on the problem of question retrieval in community Question Answering (cQA) which aims to retrieve from the community archives the previous questions that are semantically equivalent to the new queries. The major challenges in this crucial task are the shortness of the questio...

Descripción completa

Detalles Bibliográficos
Autores principales: Othman, Nouha, Faiz, Rim, Smaïli, Kamel
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7298191/
http://dx.doi.org/10.1007/978-3-030-51310-8_23
_version_ 1783547166200954880
author Othman, Nouha
Faiz, Rim
Smaïli, Kamel
author_facet Othman, Nouha
Faiz, Rim
Smaïli, Kamel
author_sort Othman, Nouha
collection PubMed
description In this paper, we focus on the problem of question retrieval in community Question Answering (cQA) which aims to retrieve from the community archives the previous questions that are semantically equivalent to the new queries. The major challenges in this crucial task are the shortness of the questions as well as the word mismatch problem as users can formulate the same query using different wording. While numerous attempts have been made to address this problem, most existing methods relied on supervised models which significantly depend on large training data sets and manual feature engineering. Such methods are mostly constrained by their specificities that put aside the word order and ignore syntactic and semantic relationships. In this work, we rely on Neural Networks (NNs) which can learn rich dense representations of text data and enable the prediction of the textual similarity between the community questions. We propose a deep learning approach based on a Siamese architecture with LSTM networks, augmented with an attention mechanism. We test different similarity measures to predict the semantic similarity between the community questions. Experiments conducted on real cQA data sets in English and Arabic show that the performance of question retrieval is improved as compared to other competitive methods.
format Online
Article
Text
id pubmed-7298191
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-72981912020-06-17 Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM Othman, Nouha Faiz, Rim Smaïli, Kamel Natural Language Processing and Information Systems Article In this paper, we focus on the problem of question retrieval in community Question Answering (cQA) which aims to retrieve from the community archives the previous questions that are semantically equivalent to the new queries. The major challenges in this crucial task are the shortness of the questions as well as the word mismatch problem as users can formulate the same query using different wording. While numerous attempts have been made to address this problem, most existing methods relied on supervised models which significantly depend on large training data sets and manual feature engineering. Such methods are mostly constrained by their specificities that put aside the word order and ignore syntactic and semantic relationships. In this work, we rely on Neural Networks (NNs) which can learn rich dense representations of text data and enable the prediction of the textual similarity between the community questions. We propose a deep learning approach based on a Siamese architecture with LSTM networks, augmented with an attention mechanism. We test different similarity measures to predict the semantic similarity between the community questions. Experiments conducted on real cQA data sets in English and Arabic show that the performance of question retrieval is improved as compared to other competitive methods. 2020-05-26 /pmc/articles/PMC7298191/ http://dx.doi.org/10.1007/978-3-030-51310-8_23 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Othman, Nouha
Faiz, Rim
Smaïli, Kamel
Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
title Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
title_full Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
title_fullStr Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
title_full_unstemmed Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
title_short Improving the Community Question Retrieval Performance Using Attention-Based Siamese LSTM
title_sort improving the community question retrieval performance using attention-based siamese lstm
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7298191/
http://dx.doi.org/10.1007/978-3-030-51310-8_23
work_keys_str_mv AT othmannouha improvingthecommunityquestionretrievalperformanceusingattentionbasedsiameselstm
AT faizrim improvingthecommunityquestionretrievalperformanceusingattentionbasedsiameselstm
AT smailikamel improvingthecommunityquestionretrievalperformanceusingattentionbasedsiameselstm