يتم استخدام AutoNCoders Varitional (VAES) على نطاق واسع للنمذجة المتغيرة الكامنة للنص.نركز على الاختلافات التي تتعلم توزيعات مسبقة معبرة على المتغير الكامن.نجد أن استراتيجيات التدريب الحالية ليست فعالة لتعلم البثور الغابات، لذلك نقترح أن نقترح إضافة احتمال هامشي لسجل الأهمية كشرطة ثانية إلى هدف VAE القياسي للمساعدة عند تعلم المقيم السابق.يؤدي القيام بذلك إلى تحسين النتائج لجميع البثور التي قامت بتقييمها، بما في ذلك اختيار جديد للجملة VAES بناء على تطبيع التدفقات (NF).لم تعد Priors المعلمة مع NF مقيدة لعائلة توزيع محددة، مما يتيح طريقة أكثر مرونة لترميز توزيع البيانات.يظهر نموذجنا، الذي نسميه FOLPRIOR، تحسنا كبيرا في مهام نمذجة اللغة مقارنة مع خطوط الأساس القوية.نحن نوضح أن flowprior يتعلم التعبير قبل التحليل والعديد من أشكال التقييم التي تنطوي على جيل.
Variational autoencoders (VAEs) are widely used for latent variable modeling of text. We focus on variations that learn expressive prior distributions over the latent variable. We find that existing training strategies are not effective for learning rich priors, so we propose adding the importance-sampled log marginal likelihood as a second term to the standard VAE objective to help when learning the prior. Doing so improves results for all priors evaluated, including a novel choice for sentence VAEs based on normalizing flows (NF). Priors parameterized with NF are no longer constrained to a specific distribution family, allowing a more flexible way to encode the data distribution. Our model, which we call FlowPrior, shows a substantial improvement in language modeling tasks compared to strong baselines. We demonstrate that FlowPrior learns an expressive prior with analysis and several forms of evaluation involving generation.
References used
https://aclanthology.org/
Undirected neural sequence models have achieved performance competitive with the state-of-the-art directed sequence models that generate monotonically from left to right in machine translation tasks. In this work, we train a policy that learns the ge
Compositional, structured models are appealing because they explicitly decompose problems and provide interpretable intermediate outputs that give confidence that the model is not simply latching onto data artifacts. Learning these models is challeng
We propose a multi-task, probabilistic approach to facilitate distantly supervised relation extraction by bringing closer the representations of sentences that contain the same Knowledge Base pairs. To achieve this, we bias the latent space of senten
The researcher in the history of ancient Eastern civilizations observes an essential aspect in them, which is the important role played by the belief in those civilizations, as it motivation and directed the foundation to her, and what submitted of m
Explaining neural network models is important for increasing their trustworthiness in real-world applications. Most existing methods generate post-hoc explanations for neural network models by identifying individual feature attributions or detecting