Cargando…

On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers

Machine reading comprehension (MRC) is one of the most challenging tasks and active fields in natural language processing (NLP). MRC systems aim to enable a machine to understand a given context in natural language and to answer a series of questions about it. With the advent of bi-directional deep...

Descripción completa

Detalles Bibliográficos
Autores principales: Ahmed, Muzamil, Khan, Hikmat, Iqbal, Tassawar, Khaled Alarfaj, Fawaz, Alomair, Abdullah, Almusallam, Naif
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10403182/
https://www.ncbi.nlm.nih.gov/pubmed/37547420
http://dx.doi.org/10.7717/peerj-cs.1422
_version_ 1785085012439203840
author Ahmed, Muzamil
Khan, Hikmat
Iqbal, Tassawar
Khaled Alarfaj, Fawaz
Alomair, Abdullah
Almusallam, Naif
author_facet Ahmed, Muzamil
Khan, Hikmat
Iqbal, Tassawar
Khaled Alarfaj, Fawaz
Alomair, Abdullah
Almusallam, Naif
author_sort Ahmed, Muzamil
collection PubMed
description Machine reading comprehension (MRC) is one of the most challenging tasks and active fields in natural language processing (NLP). MRC systems aim to enable a machine to understand a given context in natural language and to answer a series of questions about it. With the advent of bi-directional deep learning algorithms and large-scale datasets, MRC achieved improved results. However, these models are still suffering from two research issues: textual ambiguities and semantic vagueness to comprehend the long passages and generate answers for abstractive MRC systems. To address these issues, this paper proposes a novel Extended Generative Pretrained Transformers-based Question Answering (ExtGPT-QA) model to generate precise and relevant answers to questions about a given context. The proposed architecture comprises two modified forms of encoder and decoder as compared to GPT. The encoder uses a positional encoder to assign a unique representation with each word in the sentence for reference to address the textual ambiguities. Subsequently, the decoder module involves a multi-head attention mechanism along with affine and aggregation layers to mitigate semantic vagueness with MRC systems. Additionally, we applied syntax and semantic feature engineering techniques to enhance the effectiveness of the proposed model. To validate the proposed model’s effectiveness, a comprehensive empirical analysis is carried out using three benchmark datasets including SQuAD, Wiki-QA, and News-QA. The results of the proposed ExtGPT-QA outperformed state of art MRC techniques and obtained 93.25% and 90.52% F1-score and exact match, respectively. The results confirm the effectiveness of the ExtGPT-QA model to address textual ambiguities and semantic vagueness issues in MRC systems.
format Online
Article
Text
id pubmed-10403182
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-104031822023-08-05 On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers Ahmed, Muzamil Khan, Hikmat Iqbal, Tassawar Khaled Alarfaj, Fawaz Alomair, Abdullah Almusallam, Naif PeerJ Comput Sci Artificial Intelligence Machine reading comprehension (MRC) is one of the most challenging tasks and active fields in natural language processing (NLP). MRC systems aim to enable a machine to understand a given context in natural language and to answer a series of questions about it. With the advent of bi-directional deep learning algorithms and large-scale datasets, MRC achieved improved results. However, these models are still suffering from two research issues: textual ambiguities and semantic vagueness to comprehend the long passages and generate answers for abstractive MRC systems. To address these issues, this paper proposes a novel Extended Generative Pretrained Transformers-based Question Answering (ExtGPT-QA) model to generate precise and relevant answers to questions about a given context. The proposed architecture comprises two modified forms of encoder and decoder as compared to GPT. The encoder uses a positional encoder to assign a unique representation with each word in the sentence for reference to address the textual ambiguities. Subsequently, the decoder module involves a multi-head attention mechanism along with affine and aggregation layers to mitigate semantic vagueness with MRC systems. Additionally, we applied syntax and semantic feature engineering techniques to enhance the effectiveness of the proposed model. To validate the proposed model’s effectiveness, a comprehensive empirical analysis is carried out using three benchmark datasets including SQuAD, Wiki-QA, and News-QA. The results of the proposed ExtGPT-QA outperformed state of art MRC techniques and obtained 93.25% and 90.52% F1-score and exact match, respectively. The results confirm the effectiveness of the ExtGPT-QA model to address textual ambiguities and semantic vagueness issues in MRC systems. PeerJ Inc. 2023-07-24 /pmc/articles/PMC10403182/ /pubmed/37547420 http://dx.doi.org/10.7717/peerj-cs.1422 Text en ©2023 Ahmed et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Artificial Intelligence
Ahmed, Muzamil
Khan, Hikmat
Iqbal, Tassawar
Khaled Alarfaj, Fawaz
Alomair, Abdullah
Almusallam, Naif
On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
title On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
title_full On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
title_fullStr On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
title_full_unstemmed On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
title_short On solving textual ambiguities and semantic vagueness in MRC based question answering using generative pre-trained transformers
title_sort on solving textual ambiguities and semantic vagueness in mrc based question answering using generative pre-trained transformers
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10403182/
https://www.ncbi.nlm.nih.gov/pubmed/37547420
http://dx.doi.org/10.7717/peerj-cs.1422
work_keys_str_mv AT ahmedmuzamil onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers
AT khanhikmat onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers
AT iqbaltassawar onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers
AT khaledalarfajfawaz onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers
AT alomairabdullah onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers
AT almusallamnaif onsolvingtextualambiguitiesandsemanticvaguenessinmrcbasedquestionansweringusinggenerativepretrainedtransformers