إن فهم كيفية ترميز الهيكل اللغوي في التضمين السياق يمكن أن يساعد في تفسير أدائه المثير للإعجاب عبر NLP.عادة ما تدعو النهج الحالية لتحقيقها عادة إلى تدريب الطبقات وتستخدم الدقة والمعلومات المتبادلة أو التعقيد كوكيل لخير التمثيل.في هذا العمل، نجادل بأن القيام بذلك يمكن أن يكون غير موثوق به لأن تمثيلات مختلفة قد تحتاج إلى طبقات مختلفة.نقوم بتطوير إرشادي، DirectProbe، يدرس مباشرة هندسة التمثيل من خلال البناء عند فكرة مساحة الإصدار لمهمة.تبين التجارب التي لديها العديد من المهام اللغوية والموظفة السياقية أنه، حتى بدون منصوص قياسات التدريب، يمكن أن يضيء DirectProbe الأنوار حول كيفية تمثيل مساحة التضمين ملصقات وتوقع أيضا أداء المصنف للتمثيل أيضا.
Understanding how linguistic structure is encoded in contextualized embedding could help explain their impressive performance across NLP. Existing approaches for probing them usually call for training classifiers and use the accuracy, mutual information, or complexity as a proxy for the representation's goodness. In this work, we argue that doing so can be unreliable because different representations may need different classifiers. We develop a heuristic, DirectProbe, that directly studies the geometry of a representation by building upon the notion of a version space for a task. Experiments with several linguistic tasks and contextualized embeddings show that, even without training classifiers, DirectProbe can shine lights on how an embedding space represents labels and also anticipate the classifier performance for the representation.
References used
https://aclanthology.org/
This paper aims to calculate regular classical and complementary, so regular total Ignaczak solutions coupled with temperature field ,occupying R3 , and with vanishing stresses and temperature on the boundary.
Machine learning solutions are often criticized for the lack of explanation of their successes and failures. Understanding which instances are misclassified and why is essential to improve the learning process. This work helps to fill this gap by pro
We introduce SelfExplain, a novel self-explaining model that explains a text classifier's predictions using phrase-based concepts. SelfExplain augments existing neural classifiers by adding (1) a globally interpretable layer that identifies the most
Determining whether two documents were composed by the same author, also known as authorship verification, has traditionally been tackled using statistical methods. Recently, authorship representations learned using neural networks have been found to
In modern natural language processing pipelines, it is common practice to pretrain'' a generative language model on a large corpus of text, and then to finetune'' the created representations by continuing to train them on a discriminative textual inf