حققت نماذج المحولات الحديثة أداء قويا على مجموعة متنوعة من مهام NLP.توظف العديد من هذه الأساليب مهام التدريب المرجعية للمجال لتدريب النماذج التي تسفر عن تمثيلات جماعية عالية للغاية يمكن أن تكون ذات صقل مهام محددة في المصب.نقترح تكرير نموذج NLP المدرب مسبقا باستخدام هدف الكشف عن الرموز المخلوطة.نستخدم نهج متسلسل من خلال بدء تشغيل نموذج روبرتا المدرب مسبقا وتدريبه باستخدام نهجنا.تطبيق استراتيجية خلط عشوائية على مستوى الكلمة، وجدنا أن نهجنا يتيح لنموذج روبرتا يحقق أداء أفضل في 4 من أصل 7 مهام الغراء.تشير نتائجنا إلى أن تعلم الكشف عن الرموز المنفصلة هو نهج واعد لمعرفة المزيد من تمثيلات الجملة متماسكة.
State-of-the-art transformer models have achieved robust performance on a variety of NLP tasks. Many of these approaches have employed domain agnostic pre-training tasks to train models that yield highly generalized sentence representations that can be fine-tuned for specific downstream tasks. We propose refining a pre-trained NLP model using the objective of detecting shuffled tokens. We use a sequential approach by starting with the pre-trained RoBERTa model and training it using our approach. Applying random shuffling strategy on the word-level, we found that our approach enables the RoBERTa model achieve better performance on 4 out of 7 GLUE tasks. Our results indicate that learning to detect shuffled tokens is a promising approach to learn more coherent sentence representations.
References used
https://aclanthology.org/
In this paper, we describe our system used for SemEval 2021 Task 5: Toxic Spans Detection. Our proposed system approaches the problem as a token classification task. We trained our model to find toxic words and concatenate their spans to predict the
This paper presents one of the top winning solution systems for task 7 at SemEval2021, HaHackathon: Detecting and Rating Humor and Offense. This competition is divided into two tasks, task1 with three sub-tasks 1a,1b, and 1c, and task2. The goal for
In this work, we focus on a more challenging few-shot intent detection scenario where many intents are fine-grained and semantically similar. We present a simple yet effective few-shot intent detection schema via contrastive pre-training and fine-tun
Genetically modified girls are grown in many areas around the world, and the interest in planting them is increasing year after year. The area planted with genetically modified crops in 2012 amounted to about 3,170 million hectares distributed over 29 countries
Metaphors are ubiquitous in natural language, and detecting them requires contextual reasoning about whether a semantic incongruence actually exists. Most existing work addresses this problem using pre-trained contextualized models. Despite their suc