Cargando…

Reading comprehension based question answering system in Bangla language with transformer-based learning

Question answering (QA) system in any language is an assortment of mechanisms for obtaining answers to user questions with various data compositions. Reading comprehension (RC) is one type of composition, and the popularity of this type is increasing day by day in Natural Language Processing (NLP) r...

Descripción completa

Detalles Bibliográficos
Autores principales: Aurpa, Tanjim Taharat, Rifat, Richita Khandakar, Ahmed, Md Shoaib, Anwar, Md. Musfique, Ali, A. B. M. Shawkat
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9568857/
https://www.ncbi.nlm.nih.gov/pubmed/36254291
http://dx.doi.org/10.1016/j.heliyon.2022.e11052
_version_ 1784809733719326720
author Aurpa, Tanjim Taharat
Rifat, Richita Khandakar
Ahmed, Md Shoaib
Anwar, Md. Musfique
Ali, A. B. M. Shawkat
author_facet Aurpa, Tanjim Taharat
Rifat, Richita Khandakar
Ahmed, Md Shoaib
Anwar, Md. Musfique
Ali, A. B. M. Shawkat
author_sort Aurpa, Tanjim Taharat
collection PubMed
description Question answering (QA) system in any language is an assortment of mechanisms for obtaining answers to user questions with various data compositions. Reading comprehension (RC) is one type of composition, and the popularity of this type is increasing day by day in Natural Language Processing (NLP) research area. Some works have been done in several languages, mainly in English. In the Bangla language, neither any dataset available for RC nor any work has been done in the past. In this research work, we develop a question-answering system from RC. For doing this, we construct a dataset containing 3636 reading comprehensions along with questions and answers. We apply a transformer-based deep neural network model to obtain convenient answers to questions based on reading comprehensions precisely and swiftly. We exploit some deep neural network architectures such as LSTM (Long Short-Term Memory), Bi-LSTM (Bidirectional LSTM) with attention, RNN (Recurrent Neural Network), ELECTRA, and BERT (Bidirectional Encoder Representations from Transformers) to our dataset for training. The transformer-based pre-training language architectures BERT and ELECTRA perform more prominently than others from those architectures. Finally, the trained model of BERT performs a satisfactory outcome with 87.78% of testing accuracy and 99% training accuracy, and ELECTRA provides training and testing accuracy of 82.5% and 93%, respectively.
format Online
Article
Text
id pubmed-9568857
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-95688572022-10-16 Reading comprehension based question answering system in Bangla language with transformer-based learning Aurpa, Tanjim Taharat Rifat, Richita Khandakar Ahmed, Md Shoaib Anwar, Md. Musfique Ali, A. B. M. Shawkat Heliyon Research Article Question answering (QA) system in any language is an assortment of mechanisms for obtaining answers to user questions with various data compositions. Reading comprehension (RC) is one type of composition, and the popularity of this type is increasing day by day in Natural Language Processing (NLP) research area. Some works have been done in several languages, mainly in English. In the Bangla language, neither any dataset available for RC nor any work has been done in the past. In this research work, we develop a question-answering system from RC. For doing this, we construct a dataset containing 3636 reading comprehensions along with questions and answers. We apply a transformer-based deep neural network model to obtain convenient answers to questions based on reading comprehensions precisely and swiftly. We exploit some deep neural network architectures such as LSTM (Long Short-Term Memory), Bi-LSTM (Bidirectional LSTM) with attention, RNN (Recurrent Neural Network), ELECTRA, and BERT (Bidirectional Encoder Representations from Transformers) to our dataset for training. The transformer-based pre-training language architectures BERT and ELECTRA perform more prominently than others from those architectures. Finally, the trained model of BERT performs a satisfactory outcome with 87.78% of testing accuracy and 99% training accuracy, and ELECTRA provides training and testing accuracy of 82.5% and 93%, respectively. Elsevier 2022-10-12 /pmc/articles/PMC9568857/ /pubmed/36254291 http://dx.doi.org/10.1016/j.heliyon.2022.e11052 Text en © 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
spellingShingle Research Article
Aurpa, Tanjim Taharat
Rifat, Richita Khandakar
Ahmed, Md Shoaib
Anwar, Md. Musfique
Ali, A. B. M. Shawkat
Reading comprehension based question answering system in Bangla language with transformer-based learning
title Reading comprehension based question answering system in Bangla language with transformer-based learning
title_full Reading comprehension based question answering system in Bangla language with transformer-based learning
title_fullStr Reading comprehension based question answering system in Bangla language with transformer-based learning
title_full_unstemmed Reading comprehension based question answering system in Bangla language with transformer-based learning
title_short Reading comprehension based question answering system in Bangla language with transformer-based learning
title_sort reading comprehension based question answering system in bangla language with transformer-based learning
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9568857/
https://www.ncbi.nlm.nih.gov/pubmed/36254291
http://dx.doi.org/10.1016/j.heliyon.2022.e11052
work_keys_str_mv AT aurpatanjimtaharat readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning
AT rifatrichitakhandakar readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning
AT ahmedmdshoaib readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning
AT anwarmdmusfique readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning
AT aliabmshawkat readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning