يفترض تكيف المجال أن العينات من المجالات المصدر والمستهدفة يمكن الوصول إليها بحرية خلال مرحلة التدريب.ومع ذلك، نادرا ما يكون مثل هذا الافتراض معقول في العالم الحقيقي وقد يؤدي إلى مشكلات خصوصية البيانات، خاصة عندما تكون تسمية مجال المصدر يمكن أن تكون سمة حساسة كمعرف.مهمة Semeval-2021 تركز 10 على هذه القضايا.نشارك في المهمة واقتراح أطر جديدة بناء على طريقة التدريب الذاتي.في أنظمتنا، تم تصميم أطرتين مختلفتين لحل تصنيف النص ووضع التسلسل.يتم اختبار هذه الأساليب لتكون فعالة والتي تحتل المرتبة الثالثة من بين جميع النظام في التراكب الفرعي، وتحتل المرتبة الأولى بين جميع النظام في SubTask B.
Domain adaptation assumes that samples from source and target domains are freely accessible during a training phase. However, such assumption is rarely plausible in the real-world and may causes data-privacy issues, especially when the label of the source domain can be a sensitive attribute as an identifier. SemEval-2021 task 10 focuses on these issues. We participate in the task and propose novel frameworks based on self-training method. In our systems, two different frameworks are designed to solve text classification and sequence labeling. These approaches are tested to be effective which ranks the third among all system in subtask A, and ranks the first among all system in subtask B.
References used
https://aclanthology.org/
This paper presents the Source-Free Domain Adaptation shared task held within SemEval-2021. The aim of the task was to explore adaptation of machine-learning models in the face of data sharing constraints. Specifically, we consider the scenario where
Source-free domain adaptation is an emerging line of work in deep learning research since it is closely related to the real-world environment. We study the domain adaption in the sequence labeling problem where the model trained on the source domain
This paper describes our systems for negation detection and time expression recognition in SemEval 2021 Task 10, Source-Free Domain Adaptation for Semantic Processing. We show that self-training, active learning and data augmentation techniques can i
Data sharing restrictions are common in NLP datasets. The purpose of this task is to develop a model trained in a source domain to make predictions for a target domain with related domain data. To address the issue, the organizers provided the models
Due to the increasing concerns for data privacy, source-free unsupervised domain adaptation attracts more and more research attention, where only a trained source model is assumed to be available, while the labeled source data remain private. To get