تصف هذه الورقة تقديمها لمهمة LongsUMM في SDP 2021. نقترح طريقة لإدماج مظاهرة الجملة التي تنتجها نماذج لغة عميقة في تقنيات تلخيص الاستخراج بناء على مركزية الرسم البياني بطريقة غير منشأة. الطريقة المقترحة بسيطة، سريعة، يمكن أن تلخيص أينوع من وثيقة أي حجم ويمكن أن تلبي أي قيود طول الملخصات المنتجة.توفر الطريقة أداء تنافسي أساليب أكثر تطورا أكثر تطورا ويمكن أن تكون بمثابة وكيل لتقنيات تلخيص الجماع
This paper describes our submission for the LongSumm task in SDP 2021. We propose a method for incorporating sentence embeddings produced by deep language models into extractive summarization techniques based on graph centrality in an unsupervised manner.The proposed method is simple, fast, can summarize any kind of document of any size and can satisfy any length constraints for the summaries produced. The method offers competitive performance to more sophisticated supervised methods and can serve as a proxy for abstractive summarization techniques
References used
https://aclanthology.org/
Most of existing extractive multi-document summarization (MDS) methods score each sentence individually and extract salient sentences one by one to compose a summary, which have two main drawbacks: (1) neglecting both the intra and cross-document rel
Information overload has been one of the challenges regarding information from the Internet. It is not a matter of information access, instead, the focus had shifted towards the quality of the retrieved data. Particularly in the news domain, multiple
Pre-trained language models have achieved huge success on a wide range of NLP tasks. However, contextual representations from pre-trained models contain entangled semantic and syntactic information, and therefore cannot be directly used to derive use
Sentence fusion is a conditional generation task that merges several related sentences into a coherent one, which can be deemed as a summary sentence. The importance of sentence fusion has long been recognized by communities in natural language gener
Sentence extractive summarization shortens a document by selecting sentences for a summary while preserving its important contents. However, constructing a coherent and informative summary is difficult using a pre-trained BERT-based encoder since it