نستكشف استخدام مصنفات التدريب الذاتي والقبول مع النماذج المدربة مسبقا لتوليد اللغة الطبيعية في إعدادات هيكل إلى نص باستخدام ثلاث مجموعات بيانات GEM (E2E و WebNLG-EN و Schema-furdided).مع Dataset الحوار الموجهة للمخطط، نقوم أيضا بتجربة بما في ذلك المنعطفات المتعددة من السياق في المدخلات.نجد أن التدريب الذاتي مع مطابقة إعادة الإعمار مع تصفية مصنف القبول يمكن أن يحسن صحة دلالية، على الرغم من أن المكاسب محدودة في إعداد البيانات الكاملة.مع تكييف السياق، نجد أن بما في ذلك المنعطفات المتعددة في السياق يشجع النموذج على المحاذاة مع اختيارات كلمة المستخدم وصياغة وكذلك لتوليد المزيد من ردود متسقة ذاتية.في الإصدارات المستقبلية من تحدي GEM، نشجع إدراج مسارات قليلة لتشجيع البحث على كفاءة البيانات.
We explore the use of self-training and acceptability classifiers with pre-trained models for natural language generation in structure-to-text settings using three GEM datasets (E2E, WebNLG-en, Schema-Guided Dialog). With the Schema-Guided Dialog dataset, we also experiment with including multiple turns of context in the input. We find that self-training with reconstruction matching along with acceptability classifier filtering can improve semantic correctness, though gains are limited in the full-data setting. With context-conditioning, we find that including multiple turns in the context encourages the model to align with the user's word and phrasing choices as well as to generate more self-consistent responses. In future versions of the GEM challenge, we encourage the inclusion of few-shot tracks to encourage research on data efficiency.
References used
https://aclanthology.org/
The Shared Task on Hateful Memes is a challenge that aims at the detection of hateful content in memes by inviting the implementation of systems that understand memes, potentially by combining image and textual information. The challenge consists of
Abstract Recent approaches to data-to-text generation have adopted the very successful encoder-decoder architecture or variants thereof. These models generate text that is fluent (but often imprecise) and perform quite poorly at selecting appropriate
State-of-the-art deep neural networks require large-scale labeled training data that is often expensive to obtain or not available for many tasks. Weak supervision in the form of domain-specific rules has been shown to be useful in such settings to a
Despite their recent successes in tackling many NLP tasks, large-scale pre-trained language models do not perform as well in few-shot settings where only a handful of training examples are available. To address this shortcoming, we propose STraTA, wh
This paper studies the keyphrase generation (KG) task for scenarios where structure plays an important role. For example, a scientific publication consists of a short title and a long body, where the title can be used for de-emphasizing unimportant d