ستعلم وكلاء المحادثة المدربون على كوربورا كبيرة غير مبالين في التفاعلات البشرية أنماطا وسلوكيات محاكية فيها، والتي تشمل سلوكا هجوميا أو ساما.نقدم إطارا جديدا للإنسان والحلقة النموذجية لتقييم سمية هذه النماذج، ومقارنة مجموعة متنوعة من الأساليب الحالية في كل من حالات المستخدمين غير المصندين والموديين الذين يعرضون نقاط ضعفهم.ثم نذهب لاقتراح طريقتين روايتين لعوامل المحادثة الآمنة، إما عن طريق التدريب على البيانات من إطار عملنا الجديد للنسب في الحلقة في نظام من مرحلتين، أو "سلامة" في "الخبز"إلى نموذج الولادة نفسه.نجد تقنياتنا الجديدة هي (ط) أكثر أمانا من النماذج الحالية؛في حين أنه (2) الحفاظ على مقاييس قابلية الاستقرار مثل التجذير بالنسبة لشركة Chatbots الحديثة.على النقيض من ذلك، فإننا نعرض مشكلات سلامة خطيرة في الأنظمة القياسية الحالية مثل GPT2، حوالة، و BlenderBot.
Conversational agents trained on large unlabeled corpora of human interactions will learn patterns and mimic behaviors therein, which include offensive or otherwise toxic behavior. We introduce a new human-and-model-in-the-loop framework for evaluating the toxicity of such models, and compare a variety of existing methods in both the cases of non-adversarial and adversarial users that expose their weaknesses. We then go on to propose two novel methods for safe conversational agents, by either training on data from our new human-and-model-in-the-loop framework in a two-stage system, or ''baking-in'' safety to the generative model itself. We find our new techniques are (i) safer than existing models; while (ii) maintaining usability metrics such as engagingness relative to state-of-the-art chatbots. In contrast, we expose serious safety issues in existing standard systems like GPT2, DialoGPT, and BlenderBot.
References used
https://aclanthology.org/
Conversational machine reading (CMR) requires machines to communicate with humans through multi-turn interactions between two salient dialogue states of decision making and question generation processes. In open CMR settings, as the more realistic sc
Identifying relevant knowledge to be used in conversational systems that are grounded in long documents is critical to effective response generation. We introduce a knowledge identification model that leverages the document structure to provide dialo
Abstractive conversation summarization has received growing attention while most current state-of-the-art summarization models heavily rely on human-annotated summaries. To reduce the dependence on labeled summaries, in this work, we present a simple
Turn-level user satisfaction is one of the most important performance metrics for conversational agents. It can be used to monitor the agent's performance and provide insights about defective user experiences. While end-to-end deep learning has shown
Intelligent agents that are confronted with novel concepts in situated environments will need to ask their human teammates questions to learn about the physical world. To better understand this problem, we need data about asking questions in situated