Cargando…
Reading comprehension based question answering system in Bangla language with transformer-based learning
Question answering (QA) system in any language is an assortment of mechanisms for obtaining answers to user questions with various data compositions. Reading comprehension (RC) is one type of composition, and the popularity of this type is increasing day by day in Natural Language Processing (NLP) r...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9568857/ https://www.ncbi.nlm.nih.gov/pubmed/36254291 http://dx.doi.org/10.1016/j.heliyon.2022.e11052 |
_version_ | 1784809733719326720 |
---|---|
author | Aurpa, Tanjim Taharat Rifat, Richita Khandakar Ahmed, Md Shoaib Anwar, Md. Musfique Ali, A. B. M. Shawkat |
author_facet | Aurpa, Tanjim Taharat Rifat, Richita Khandakar Ahmed, Md Shoaib Anwar, Md. Musfique Ali, A. B. M. Shawkat |
author_sort | Aurpa, Tanjim Taharat |
collection | PubMed |
description | Question answering (QA) system in any language is an assortment of mechanisms for obtaining answers to user questions with various data compositions. Reading comprehension (RC) is one type of composition, and the popularity of this type is increasing day by day in Natural Language Processing (NLP) research area. Some works have been done in several languages, mainly in English. In the Bangla language, neither any dataset available for RC nor any work has been done in the past. In this research work, we develop a question-answering system from RC. For doing this, we construct a dataset containing 3636 reading comprehensions along with questions and answers. We apply a transformer-based deep neural network model to obtain convenient answers to questions based on reading comprehensions precisely and swiftly. We exploit some deep neural network architectures such as LSTM (Long Short-Term Memory), Bi-LSTM (Bidirectional LSTM) with attention, RNN (Recurrent Neural Network), ELECTRA, and BERT (Bidirectional Encoder Representations from Transformers) to our dataset for training. The transformer-based pre-training language architectures BERT and ELECTRA perform more prominently than others from those architectures. Finally, the trained model of BERT performs a satisfactory outcome with 87.78% of testing accuracy and 99% training accuracy, and ELECTRA provides training and testing accuracy of 82.5% and 93%, respectively. |
format | Online Article Text |
id | pubmed-9568857 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-95688572022-10-16 Reading comprehension based question answering system in Bangla language with transformer-based learning Aurpa, Tanjim Taharat Rifat, Richita Khandakar Ahmed, Md Shoaib Anwar, Md. Musfique Ali, A. B. M. Shawkat Heliyon Research Article Question answering (QA) system in any language is an assortment of mechanisms for obtaining answers to user questions with various data compositions. Reading comprehension (RC) is one type of composition, and the popularity of this type is increasing day by day in Natural Language Processing (NLP) research area. Some works have been done in several languages, mainly in English. In the Bangla language, neither any dataset available for RC nor any work has been done in the past. In this research work, we develop a question-answering system from RC. For doing this, we construct a dataset containing 3636 reading comprehensions along with questions and answers. We apply a transformer-based deep neural network model to obtain convenient answers to questions based on reading comprehensions precisely and swiftly. We exploit some deep neural network architectures such as LSTM (Long Short-Term Memory), Bi-LSTM (Bidirectional LSTM) with attention, RNN (Recurrent Neural Network), ELECTRA, and BERT (Bidirectional Encoder Representations from Transformers) to our dataset for training. The transformer-based pre-training language architectures BERT and ELECTRA perform more prominently than others from those architectures. Finally, the trained model of BERT performs a satisfactory outcome with 87.78% of testing accuracy and 99% training accuracy, and ELECTRA provides training and testing accuracy of 82.5% and 93%, respectively. Elsevier 2022-10-12 /pmc/articles/PMC9568857/ /pubmed/36254291 http://dx.doi.org/10.1016/j.heliyon.2022.e11052 Text en © 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Research Article Aurpa, Tanjim Taharat Rifat, Richita Khandakar Ahmed, Md Shoaib Anwar, Md. Musfique Ali, A. B. M. Shawkat Reading comprehension based question answering system in Bangla language with transformer-based learning |
title | Reading comprehension based question answering system in Bangla language with transformer-based learning |
title_full | Reading comprehension based question answering system in Bangla language with transformer-based learning |
title_fullStr | Reading comprehension based question answering system in Bangla language with transformer-based learning |
title_full_unstemmed | Reading comprehension based question answering system in Bangla language with transformer-based learning |
title_short | Reading comprehension based question answering system in Bangla language with transformer-based learning |
title_sort | reading comprehension based question answering system in bangla language with transformer-based learning |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9568857/ https://www.ncbi.nlm.nih.gov/pubmed/36254291 http://dx.doi.org/10.1016/j.heliyon.2022.e11052 |
work_keys_str_mv | AT aurpatanjimtaharat readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning AT rifatrichitakhandakar readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning AT ahmedmdshoaib readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning AT anwarmdmusfique readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning AT aliabmshawkat readingcomprehensionbasedquestionansweringsysteminbanglalanguagewithtransformerbasedlearning |