توضيحات تناقض توضيح سبب حدوث حدث قد حدث على عكس آخر.إنهم بطبيعتهم بديهية للبشر لكل من الإنتاج والفهم.نقترح طريقة لإنتاج تفسيرات صغيرة في الفضاء الكامن، من خلال إسقاط تمثيل الإدخال، بحيث يتم التقاط الميزات التي تفرق إلا عن قرارات محتملة.يسمح التعديل لدينا بسلوك نموذجي للنظر في التفكير المتعرج فقط، والكشف عن جوانب المدخلات مفيدة لقرارات ومعاكضة معينة.يمكن أن تجيب تفسيراتنا الصنع للإجابة على أي تسمية، ومعها الملصق البدائل، هي ميزة إدخال معينة مفيدة.نحن ننتج تفسيرات صغيرة عبر كل من إسناد مفهوم مجردة رفيع المستوى ومستوى المدخلات / المدخلات منخفضة المستوى لإسناد معايير تصنيف NLP.توضح نتائجنا قدرة التفسيرات على نطاق واسع لتوفير إمكانية الترجمة الترجمة الفورية للقرارات النموذجية.
Contrastive explanations clarify why an event occurred in contrast to another. They are inherently intuitive to humans to both produce and comprehend. We propose a method to produce contrastive explanations in the latent space, via a projection of the input representation, such that only the features that differentiate two potential decisions are captured. Our modification allows model behavior to consider only contrastive reasoning, and uncover which aspects of the input are useful for and against particular decisions. Our contrastive explanations can additionally answer for which label, and against which alternative label, is a given input feature useful. We produce contrastive explanations via both high-level abstract concept attribution and low-level input token/span attribution for two NLP classification benchmarks. Our findings demonstrate the ability of label-contrastive explanations to provide fine-grained interpretability of model decisions.
References used
https://aclanthology.org/
Ideally, people who navigate together in a complex indoor space share a mental model that facilitates explanation. This paper reports on a robot control system whose cognitive world model is based on spatial affordances that generalize over its perce
In this paper, we explore the construction of natural language explanations for news claims, with the goal of assisting fact-checking and news evaluation applications. We experiment with two methods: (1) an extractive method based on Biased TextRank
How can we generate concise explanations for multi-hop Reading Comprehension (RC)? The current strategies of identifying supporting sentences can be seen as an extractive question-focused summarization of the input text. However, these extractive exp
Although neural models have shown strong performance in datasets such as SNLI, they lack the ability to generalize out-of-distribution (OOD). In this work, we formulate a few-shot learning setup and examine the effects of natural language explanation
Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the