نقدم أول محلل محلل دائري على أساس أنظمة إعادة الكتابة الخالية من السياق (LCFRS).وهو يستخدم المصنفات العصبية والتفوق على المحللين السابقين في LCFRS في كل من الدقة وسرعة التحليل من هامش واسع.نتائجنا مواكبة أفضل المحللين المتساقين (العام)، وخاصة درجات الناخبين المتساقين إنشاء حالة جديدة من الفن.إن قلب نهجنا هو إجراء فعال لليككاليزي يدفع LCFRS معجمية من أي شجرة Treebank غير المتساقين.وصفنا بتعديل تحليل LCFRS المعتاد على الرسم البياني الذي يمثل التضخم بسرعة وإدخال إجراء يحول مشتقات LCFRS المعجمية إلى أشجار تحليل مكافئة من TreeBank الأصلي.يتم تقييم نهجنا على الانجليزية المتساقين بين بنك بنك بنسل وشركة النيجرية الألمانية والنمر.
We present the first supertagging-based parser for linear context-free rewriting systems (LCFRS). It utilizes neural classifiers and outperforms previous LCFRS-based parsers in both accuracy and parsing speed by a wide margin. Our results keep up with the best (general) discontinuous parsers, particularly the scores for discontinuous constituents establish a new state of the art. The heart of our approach is an efficient lexicalization procedure which induces a lexical LCFRS from any discontinuous treebank. We describe a modification to usual chart-based LCFRS parsing that accounts for supertagging and introduce a procedure that transforms lexical LCFRS derivations into equivalent parse trees of the original treebank. Our approach is evaluated on the English Discontinuous Penn Treebank and the German treebanks Negra and Tiger.
References used
https://aclanthology.org/
Paraphrase generation is an important task in natural language processing. Previous works focus on sentence-level paraphrase generation, while ignoring document-level paraphrase generation, which is a more challenging and valuable task. In this paper
Since their inception, transformer-based language models have led to impressive performance gains across multiple natural language processing tasks. For Arabic, the current state-of-the-art results on most datasets are achieved by the AraBERT languag
This paper details experiments we performed on the Universal Dependencies 2.7 corpora in order to investigate the dominant word order in the available languages. For this purpose, we used a graph rewriting tool, GREW, which allowed us to go beyond th
Probabilistic context-free grammars (PCFGs) with neural parameterization have been shown to be effective in unsupervised phrase-structure grammar induction. However, due to the cubic computational complexity of PCFG representation and parsing, previo
This paper describes a system proposed for the IWPT 2021 Shared Task on Parsing into Enhanced Universal Dependencies (EUD). We propose a Graph Rewriting based system for computing Enhanced Universal Dependencies, given the Basic Universal Dependencies (UD).