نقترح إطارا عاما للترجمة الآلية المتزامنة.تستخدم النهج التقليدية عددا ثابتا من الكلمات المصدر لترجمة أو تعلم السياسات الديناميكية لعدد الكلمات المصدر عن طريق التعلم التعزيز.نحن هنا صياغة ترجمة متزامنة كمشكلة تعلم التسلسل الهيكلية إلى التسلسل.يتم تقديم متغير كامن إلى نموذج قراءة أو ترجمة الإجراءات في كل خطوة زمنية، ثم يتم دمجها بعد ذلك للنظر في جميع سياسات الترجمة الممكنة.يستخدم POISSON RE-PLISTIONSED قبل تنظيم السياسات التي تسمح للنموذج بتوازن بشكل صريح بجودة الترجمة والكمول.توضح التجارب فعالية وأغاني الإطار الإداري، والذي يحقق أفضل درجات بلو نظرا لمتوسط الألوان المتوسطة عن مصطلحات البيانات القياسية.
We propose a generative framework for simultaneous machine translation. Conventional approaches use a fixed number of source words to translate or learn dynamic policies for the number of source words by reinforcement learning. Here we formulate simultaneous translation as a structural sequence-to-sequence learning problem. A latent variable is introduced to model read or translate actions at every time step, which is then integrated out to consider all the possible translation policies. A re-parameterised Poisson prior is used to regularise the policies which allows the model to explicitly balance translation quality and latency. The experiments demonstrate the effectiveness and robustness of the generative framework, which achieves the best BLEU scores given different average translation latencies on benchmark datasets.
References used
https://aclanthology.org/
Simultaneous machine translation has recently gained traction thanks to significant quality improvements and the advent of streaming applications. Simultaneous translation systems need to find a trade-off between translation quality and response time
Most existing simultaneous machine translation (SiMT) systems are trained and evaluated on offline translation corpora. We argue that SiMT systems should be trained and tested on real interpretation data. To illustrate this argument, we propose an in
Neural machine translation (NMT) models are data-driven and require large-scale training corpus. In practical applications, NMT models are usually trained on a general domain corpus and then fine-tuned by continuing training on the in-domain corpus.
Recent research questions the importance of the dot-product self-attention in Transformer models and shows that most attention heads learn simple positional patterns. In this paper, we push further in this research line and propose a novel substitute
There are common semantics shared across text and images. Given a sentence in a source language, whether depicting the visual scene helps translation into a target language? Existing multimodal neural machine translation methods (MNMT) require triple