حظي مؤخرا اختصاص البيانات الضخمة باهتمام كبير في مجالات متنوعة منها (الطب , العلوم , الادارة, السياسة , ......)
و يهتم هذا الاختصاص بدراسة مجموعة البيانات الضخمة والتي تعجز الادوات والطرق الشائعة على معالجتها و ادارتها و تنظيمها خلال فترة زمنية مقبولة و بناء نموذج للتعامل مع هذه المعطيات والتنبؤ باغراض مطلوبة منها.
ولاجراء هذه الدراسات ظهرت طرق عدة منها النماذج التي تعتمد على مجموعة من البيانات و نماذج تعتمد على المحاكاة و في هذه المقالة تم توضيح الفرق بين النموذجين و تطبيق نهج جديد يعتمد على التكامل بين النموذجين لاعطاء نموذح افضل لمعالجة مسالة البيوت البلاستيكة
No English abstract
In recent years, time-critical processing or real-time processing and analytics of bid data have received a significant amount of attentions. There are many areas/domains where real-time processing of data and making timely decision can save thousand
Through our study, the HadoopOperationTesting software library
has been developed to provide Big Data applications labs with a
mechanism to test their applications in a simulated environment for
the Hadoop environment with a similar mechanism to test
traditional applications using the JUnit library.
Pre-trained neural language models give high performance on natural language inference (NLI) tasks. But whether they actually understand the meaning of the processed sequences is still unclear. We propose a new diagnostics test suite which allows to
Developing documentation guidelines and easy-to-use templates for datasets and models is a challenging task, especially given the variety of backgrounds, skills, and incentives of the people involved in the building of natural language processing (NL
We address the problem of language model customization in applications where the ASR component needs to manage domain-specific terminology; although current state-of-the-art speech recognition technology provides excellent results for generic domains