على الرغم من أن النماذج العصبية قد أظهرت أداء قويا في مجموعات البيانات مثل SNLI، إلا أنها تفتقر إلى القدرة على التعميم خارج التوزيع (OOD).في هذا العمل، نقوم بصياغة عدد قليل من إعدادات التعلم ودراسة آثار تفسيرات اللغة الطبيعية على تعميم OOD.نحن نستفيد من القوالب في DataSet Hans وبناء تفسيرات لغة طبيعية TEMPLated لكل قالب.على الرغم من أن التفسيرات الناتجة تظهر درجات بلو تنافسية ضد تفسيرات الحقيقة الأرضية، إلا أنها تفشل في تحسين أداء التنبؤ.نوضح مزيد من التفسيرات التي تم إنشاؤها في كثير من الأحيان معلومات الهلوسة والأيس عن العناصر الرئيسية التي تشير إلى الملصق.
Although neural models have shown strong performance in datasets such as SNLI, they lack the ability to generalize out-of-distribution (OOD). In this work, we formulate a few-shot learning setup and examine the effects of natural language explanations on OOD generalization. We leverage the templates in the HANS dataset and construct templated natural language explanations for each template. Although generated explanations show competitive BLEU scores against ground truth explanations, they fail to improve prediction performance. We further show that generated explanations often hallucinate information and miss key elements that indicate the label.
References used
https://aclanthology.org/
An emerging line of research in Explainable NLP is the creation of datasets enriched with human-annotated explanations and rationales, used to build and evaluate models with step-wise inference and explanation generation capabilities. While human-ann
Much of recent progress in NLU was shown to be due to models' learning dataset-specific heuristics. We conduct a case study of generalization in NLI (from MNLI to the adversarially constructed HANS dataset) in a range of BERT-based architectures (ada
Providing pretrained language models with simple task descriptions in natural language enables them to solve some tasks in a fully unsupervised fashion. Moreover, when combined with regular learning from examples, this idea yields impressive few-shot
In this paper, we study the utilization of pre-trained language models to enable few-shotNatural Language Generation (NLG) in task-oriented dialog systems. We introduce a system consisting of iterative self-training and an extensible mini-template fr
Pre-trained neural language models give high performance on natural language inference (NLI) tasks. But whether they actually understand the meaning of the processed sequences is still unclear. We propose a new diagnostics test suite which allows to