Cargando…
On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
Machine reading comprehension (MRC) is one of the most challenging tasks and active fields in natural language processing (NLP). MRC systems aim to enable a machine to understand a given context in natural language and to answer a series of questions about it. With the advent of bi-directional deep...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
PeerJ Inc.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10403182/ https://www.ncbi.nlm.nih.gov/pubmed/37547420 http://dx.doi.org/10.7717/peerj-cs.1422 |
_version_ | 1785085012439203840 |
---|---|
author | Ahmed, Muzamil Khan, Hikmat Iqbal, Tassawar Khaled Alarfaj, Fawaz Alomair, Abdullah Almusallam, Naif |
author_facet | Ahmed, Muzamil Khan, Hikmat Iqbal, Tassawar Khaled Alarfaj, Fawaz Alomair, Abdullah Almusallam, Naif |
author_sort | Ahmed, Muzamil |
collection | PubMed |
description | Machine reading comprehension (MRC) is one of the most challenging tasks and active fields in natural language processing (NLP). MRC systems aim to enable a machine to understand a given context in natural language and to answer a series of questions about it. With the advent of bi-directional deep learning algorithms and large-scale datasets, MRC achieved improved results. However, these models are still suffering from two research issues: textual ambiguities and semantic vagueness to comprehend the long passages and generate answers for abstractive MRC systems. To address these issues, this paper proposes a novel Extended Generative Pretrained Transformers-based Question Answering (ExtGPT-QA) model to generate precise and relevant answers to questions about a given context. The proposed architecture comprises two modified forms of encoder and decoder as compared to GPT. The encoder uses a positional encoder to assign a unique representation with each word in the sentence for reference to address the textual ambiguities. Subsequently, the decoder module involves a multi-head attention mechanism along with affine and aggregation layers to mitigate semantic vagueness with MRC systems. Additionally, we applied syntax and semantic feature engineering techniques to enhance the effectiveness of the proposed model. To validate the proposed model’s effectiveness, a comprehensive empirical analysis is carried out using three benchmark datasets including SQuAD, Wiki-QA, and News-QA. The results of the proposed ExtGPT-QA outperformed state of art MRC techniques and obtained 93.25% and 90.52% F1-score and exact match, respectively. The results confirm the effectiveness of the ExtGPT-QA model to address textual ambiguities and semantic vagueness issues in MRC systems. |
format | Online Article Text |
id | pubmed-10403182 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | PeerJ Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-104031822023-08-05 On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers Ahmed, Muzamil Khan, Hikmat Iqbal, Tassawar Khaled Alarfaj, Fawaz Alomair, Abdullah Almusallam, Naif PeerJ Comput Sci Artificial Intelligence Machine reading comprehension (MRC) is one of the most challenging tasks and active fields in natural language processing (NLP). MRC systems aim to enable a machine to understand a given context in natural language and to answer a series of questions about it. With the advent of bi-directional deep learning algorithms and large-scale datasets, MRC achieved improved results. However, these models are still suffering from two research issues: textual ambiguities and semantic vagueness to comprehend the long passages and generate answers for abstractive MRC systems. To address these issues, this paper proposes a novel Extended Generative Pretrained Transformers-based Question Answering (ExtGPT-QA) model to generate precise and relevant answers to questions about a given context. The proposed architecture comprises two modified forms of encoder and decoder as compared to GPT. The encoder uses a positional encoder to assign a unique representation with each word in the sentence for reference to address the textual ambiguities. Subsequently, the decoder module involves a multi-head attention mechanism along with affine and aggregation layers to mitigate semantic vagueness with MRC systems. Additionally, we applied syntax and semantic feature engineering techniques to enhance the effectiveness of the proposed model. To validate the proposed model’s effectiveness, a comprehensive empirical analysis is carried out using three benchmark datasets including SQuAD, Wiki-QA, and News-QA. The results of the proposed ExtGPT-QA outperformed state of art MRC techniques and obtained 93.25% and 90.52% F1-score and exact match, respectively. The results confirm the effectiveness of the ExtGPT-QA model to address textual ambiguities and semantic vagueness issues in MRC systems. PeerJ Inc. 2023-07-24 /pmc/articles/PMC10403182/ /pubmed/37547420 http://dx.doi.org/10.7717/peerj-cs.1422 Text en ©2023 Ahmed et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
spellingShingle | Artificial Intelligence Ahmed, Muzamil Khan, Hikmat Iqbal, Tassawar Khaled Alarfaj, Fawaz Alomair, Abdullah Almusallam, Naif On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers |
title | On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers |
title_full | On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers |
title_fullStr | On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers |
title_full_unstemmed | On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers |
title_short | On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers |
title_sort | on solving textual ambiguities and semantic vagueness in mrc based question answering using generative pre-trained transformers |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10403182/ https://www.ncbi.nlm.nih.gov/pubmed/37547420 http://dx.doi.org/10.7717/peerj-cs.1422 |
work_keys_str_mv | AT ahmedmuzamil onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers AT khanhikmat onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers AT iqbaltassawar onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers AT khaledalarfajfawaz onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers AT alomairabdullah onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers AT almusallamnaif onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers |