Cargando…

Ask me in your own words: paraphrasing for multitask question answering

Multitask learning has led to significant advances in Natural Language Processing, including the decaNLP benchmark where question answering is used to frame 10 natural language understanding tasks in a single model. In this work we show how models trained to solve decaNLP fail with simple paraphrasi...

Descripción completa

Detalles Bibliográficos
Autores principales: Hudson, G. Thomas, Al Moubayed, Noura
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8576550/
https://www.ncbi.nlm.nih.gov/pubmed/34805510
http://dx.doi.org/10.7717/peerj-cs.759
_version_ 1784595899324825600
author Hudson, G. Thomas
Al Moubayed, Noura
author_facet Hudson, G. Thomas
Al Moubayed, Noura
author_sort Hudson, G. Thomas
collection PubMed
description Multitask learning has led to significant advances in Natural Language Processing, including the decaNLP benchmark where question answering is used to frame 10 natural language understanding tasks in a single model. In this work we show how models trained to solve decaNLP fail with simple paraphrasing of the question. We contribute a crowd-sourced corpus of paraphrased questions (PQ-decaNLP), annotated with paraphrase phenomena. This enables analysis of how transformations such as swapping the class labels and changing the sentence modality lead to a large performance degradation. Training both MQAN and the newer T5 model using PQ-decaNLP improves their robustness and for some tasks improves the performance on the original questions, demonstrating the benefits of a model which is more robust to paraphrasing. Additionally, we explore how paraphrasing knowledge is transferred between tasks, with the aim of exploiting the multitask property to improve the robustness of the models. We explore the addition of paraphrase detection and paraphrase generation tasks, and find that while both models are able to learn these new tasks, knowledge about paraphrasing does not transfer to other decaNLP tasks.
format Online
Article
Text
id pubmed-8576550
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-85765502021-11-19 Ask me in your own words: paraphrasing for multitask question answering Hudson, G. Thomas Al Moubayed, Noura PeerJ Comput Sci Computational Linguistics Multitask learning has led to significant advances in Natural Language Processing, including the decaNLP benchmark where question answering is used to frame 10 natural language understanding tasks in a single model. In this work we show how models trained to solve decaNLP fail with simple paraphrasing of the question. We contribute a crowd-sourced corpus of paraphrased questions (PQ-decaNLP), annotated with paraphrase phenomena. This enables analysis of how transformations such as swapping the class labels and changing the sentence modality lead to a large performance degradation. Training both MQAN and the newer T5 model using PQ-decaNLP improves their robustness and for some tasks improves the performance on the original questions, demonstrating the benefits of a model which is more robust to paraphrasing. Additionally, we explore how paraphrasing knowledge is transferred between tasks, with the aim of exploiting the multitask property to improve the robustness of the models. We explore the addition of paraphrase detection and paraphrase generation tasks, and find that while both models are able to learn these new tasks, knowledge about paraphrasing does not transfer to other decaNLP tasks. PeerJ Inc. 2021-10-27 /pmc/articles/PMC8576550/ /pubmed/34805510 http://dx.doi.org/10.7717/peerj-cs.759 Text en © 2021 Hudson and Al Moubayed https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Computational Linguistics
Hudson, G. Thomas
Al Moubayed, Noura
Ask me in your own words: paraphrasing for multitask question answering
title Ask me in your own words: paraphrasing for multitask question answering
title_full Ask me in your own words: paraphrasing for multitask question answering
title_fullStr Ask me in your own words: paraphrasing for multitask question answering
title_full_unstemmed Ask me in your own words: paraphrasing for multitask question answering
title_short Ask me in your own words: paraphrasing for multitask question answering
title_sort ask me in your own words: paraphrasing for multitask question answering
topic Computational Linguistics
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8576550/
https://www.ncbi.nlm.nih.gov/pubmed/34805510
http://dx.doi.org/10.7717/peerj-cs.759
work_keys_str_mv AT hudsongthomas askmeinyourownwordsparaphrasingformultitaskquestionanswering
AT almoubayednoura askmeinyourownwordsparaphrasingformultitaskquestionanswering