Ask me in your own words: paraphrasing for multitask question answering.

G Thomas Hudson, Noura Al Moubayed
Author Information
  1. G Thomas Hudson: Department of Computer Science, Durham University, Durham, United Kingdom. ORCID
  2. Noura Al Moubayed: Department of Computer Science, Durham University, Durham, United Kingdom.

Abstract

Multitask learning has led to significant advances in Natural Language Processing, including the decaNLP benchmark where question answering is used to frame 10 natural language understanding tasks in a single model. In this work we show how models trained to solve decaNLP fail with simple paraphrasing of the question. We contribute a crowd-sourced of paraphrased questions (PQ-decaNLP), annotated with paraphrase phenomena. This enables analysis of how transformations such as swapping the class labels and changing the sentence modality lead to a large performance degradation. Training both MQAN and the newer T5 model using PQ-decaNLP improves their robustness and for some tasks improves the performance on the original questions, demonstrating the benefits of a model which is more robust to paraphrasing. Additionally, we explore how paraphrasing knowledge is transferred between tasks, with the aim of exploiting the multitask property to improve the robustness of the models. We explore the addition of paraphrase detection and paraphrase generation tasks, and find that while both models are able to learn these new tasks, knowledge about paraphrasing does not transfer to other decaNLP tasks.

Keywords

Word Cloud

Created with Highcharts 10.0.0tasksparaphrasingdecaNLPquestionansweringmodelmodelsparaphraseMultitasklearningquestionsPQ-decaNLPperformanceimprovesrobustnessexploreknowledgemultitaskledsignificantadvancesNaturalLanguageProcessingincludingbenchmarkusedframe10naturallanguageunderstandingsingleworkshowtrainedsolvefailsimplecontributecrowd-sourcedparaphrasedannotatedphenomenaenablesanalysistransformationsswappingclasslabelschangingsentencemodalityleadlargedegradationTrainingMQANnewerT5usingoriginaldemonstratingbenefitsrobustAdditionallytransferredaimexploitingpropertyimproveadditiondetectiongenerationfindablelearnnewtransferAskwords:DatasetParaphrasingQuestion

Similar Articles

Cited By