تتحول نماذج المحادثة واسعة النطاق إلى الاستفادة من المعرفة الخارجية لتحسين الدقة الواقعية في توليد الاستجابة.بالنظر إلى عدم التعليق على المعرفة الخارجية لعوريا الحوار واسعة النطاق، من المستحسن معرفة اختيار المعرفة وتوليد الاستجابة بطريقة غير منشأة.في هذه الورقة، نقترح أفلاطون كاج (توليد المعرفة المعزز)، ونهج تعليمي غير مخطط له لنمذجة المحادثة المحفوظة على المعرفة الطرفية.لكل سياق حوار، يتم اختيار عناصر المعرفة ذات الصلة من الأعلى وبعد ذلك في توليد الاستجابة المدرجة في المعرفة.يتم تحسين مكونين اختيار المعرفة وتوليد الاستجابة بشكل مشترك وفعال تحت هدف متوازن.النتائج التجريبية على اثنين من مجموعات البيانات المتاحة للجمهور التحقق من تفوق أفلاطون كاج.
Large-scale conversation models are turning to leveraging external knowledge to improve the factual accuracy in response generation. Considering the infeasibility to annotate the external knowledge for large-scale dialogue corpora, it is desirable to learn the knowledge selection and response generation in an unsupervised manner. In this paper, we propose PLATO-KAG (Knowledge-Augmented Generation), an unsupervised learning approach for end-to-end knowledge-grounded conversation modeling. For each dialogue context, the top-k relevant knowledge elements are selected and then employed in knowledge-grounded response generation. The two components of knowledge selection and response generation are optimized jointly and effectively under a balanced objective. Experimental results on two publicly available datasets validate the superiority of PLATO-KAG.
References used
https://aclanthology.org/
Generating informative and appropriate responses is challenging but important for building human-like dialogue systems. Although various knowledge-grounded conversation models have been proposed, these models have limitations in utilizing knowledge t
Conversation disentanglement aims to separate intermingled messages into detached sessions, which is a fundamental task in understanding multi-party conversations. Existing work on conversation disentanglement relies heavily upon human-annotated data
Knowledge data are massive and widespread in the real-world, which can serve as good external sources to enrich conversations. However, in knowledge-grounded conversations, current models still lack the fine-grained control over knowledge selection a
We present Graformer, a novel Transformer-based encoder-decoder architecture for graph-to-text generation. With our novel graph self-attention, the encoding of a node relies on all nodes in the input graph - not only direct neighbors - facilitating t
The nature of no word delimiter or inflection that can indicate segment boundaries or word semantics increases the difficulty of Chinese text understanding, and also intensifies the demand for word-level semantic knowledge to accomplish the tagging g