تقدم الورقة أربع نماذج مقدمة إلى الجزء 2 من المهمة المشتركة Sigmorphon 2021 0، التي تهدف إلى تكرار الأحكام الإنسانية على انعطاف أحادي الإكسآت.هدفنا هو استكشاف فائدة الجمع بين الأنماط التناظرية التي تم تجميعها مسبقا مع بنية تشفير فك الترميز.تم تصميم نموذجين باستخدام هذه الأنماط إما في الإدخال أو إخراج الشبكة.نماذج إضافية يتم التحكم فيها لدور التشابه الخام للنماذج المؤذية غير المصنفة للأشكال المصابة الموجودة في نفس خلية النموذج، ودور تواتر نوع الأنماط التناظرية.استراتيجيتنا غير داخلي تماما بمعنى أن النماذج تستأنف فقط البيانات المقدمة من منظمي Sigmorphon، دون استخدام موارد خارجية.تحتل نموذجنا 2 المرتبة الثانية بين جميع الأنظمة المقدمة، مما يشير إلى أن إدراج أنماط تكنولوجية في بنية الشبكة مفيدة في تنبؤات مكبرات الصوت المحاكمة.
The paper presents four models submitted to Part 2 of the SIGMORPHON 2021 Shared Task 0, which aims at replicating human judgements on the inflection of nonce lexemes. Our goal is to explore the usefulness of combining pre-compiled analogical patterns with an encoder-decoder architecture. Two models are designed using such patterns either in the input or the output of the network. Two extra models controlled for the role of raw similarity of nonce inflected forms to existing inflected forms in the same paradigm cell, and the role of the type frequency of analogical patterns. Our strategy is entirely endogenous in the sense that the models appealing solely to the data provided by the SIGMORPHON organisers, without using external resources. Our model 2 ranks second among all submitted systems, suggesting that the inclusion of analogical patterns in the network architecture is useful in mimicking speakers' predictions.
References used
https://aclanthology.org/
Morphological rules with various levels of specificity can be learned from example lexemes by recursive application of minimal generalization (Albright and Hayes, 2002, 2003).A model that learns rules solely through minimal generalization was used to
Encoder-decoder models have been commonly used for many tasks such as machine translation and response generation. As previous research reported, these models suffer from generating redundant repetition. In this research, we propose a new mechanism f
Due to its effectiveness and performance, the Transformer translation model has attracted wide attention, most recently in terms of probing-based approaches. Previous work focuses on using or probing source linguistic features in the encoder. To date
We often use perturbations to regularize neural models. For neural encoder-decoders, previous studies applied the scheduled sampling (Bengio et al., 2015) and adversarial perturbations (Sato et al., 2019) as perturbations but these methods require co
Curriculum learning, a machine training strategy that feeds training instances to the model from easy to hard, has been proven to facilitate the dialogue generation task. Meanwhile, knowledge distillation, a knowledge transformation methodology among