Do you want to publish a course? Click here

Improving the extraction of audio features In audio-visual Arabic systems

تحسين استخراج السمات السمعية في الأنظمة السمعية البصرية للمتحدثين باللغة العربية

1217   2   49   0 ( 0 )
 Publication date 2017
and research's language is العربية
 Created by Shamra Editor




Ask ChatGPT about the research

The audio-visual speech recognition systems that rely on speech and movement of the lips of the speaker of the most important speech recognition systems. Many different techniques have developed in terms of the methods used in the feature extraction and classification methods. Research proposes the establishment of a system to identify isolated words based audio features extracted from videos pronunciations of words in Arabic in an environment free of noise, and then add the energy and Temporal derivative components in extracting features of the method Mel Frequency Cepstral Coefficient (MFCC) stage.

References used
Marius Zbancioc, Mihaela Costin :using neural networks and LPCC to improve speech recognition, International IEEE SCS Conference, Proceedings, Vol. 1, 2003 EX 720, pp. 445 – 448
Levy, C., Linares, G., Nocera, P., Bonastre, J.-F. : Reducing computational and memory cost for cellular phone embedded speech recognition system, Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on (Volume:5 ) , pages( 309-12) vol.5 , Print ISBN:9-8484-7803-0
Dimitriadis, Maragos, P. Potamianos: Robust AM-FM Features for Speech Recognition, IEEE signal processing letters, VOL. 12, NO. 9, 2005
rate research

Read More

This study aims to score psychological, academic and economic problems faced by students with special needs in the Faculty of Arts at the University of Damascus and to identify the differences in these problems between the research samples. The sa mple consisted of / 72/ students from the students of the Faculty of Arts at Damascus University, was selected deliberate manner.
we'll show in this research months fingerprints developed in the criminal investigation, a visual and sound footprint, and made it clear that these fingerprints applications of interest to specialists in the science crime and criminal evidence, it is through the visual and sound footprint can be identified by the accused and convicted of their crime or brought them, and finally we determined that This kind of fingerprinting subject to the principle of legality of evidence like any other evidence, and nothing prevents the use in evidence penal taking into account some of the controls that are indispensable to Legitimacy.
Story visualization is an underexplored task that falls at the intersection of many important research directions in both computer vision and natural language processing. In this task, given a series of natural language captions which compose a story , an agent must generate a sequence of images that correspond to the captions. Prior work has introduced recurrent generative models which outperform text-to-image synthesis models on this task. However, there is room for improvement of generated images in terms of visual quality, coherence and relevance. We present a number of improvements to prior modeling approaches, including (1) the addition of a dual learning framework that utilizes video captioning to reinforce the semantic alignment between the story and generated images, (2) a copy-transform mechanism for sequentially-consistent story visualization, and (3) MART-based transformers to model complex interactions between frames. We present ablation studies to demonstrate the effect of each of these techniques on the generative power of the model for both individual images as well as the entire narrative. Furthermore, due to the complexity and generative nature of the task, standard evaluation metrics do not accurately reflect performance. Therefore, we also provide an exploration of evaluation metrics for the model, focused on aspects of the generated frames such as the presence/quality of generated characters, the relevance to captions, and the diversity of the generated images. We also present correlation experiments of our proposed automated metrics with human evaluations.
In recent years, the emergence of streaming platforms such as Netflix, HBO or Amazon Prime Video has reshaped the field of entertainment, which increasingly relies on subtitling, dubbing or voice-over modes. However, little is known about audiovisual translation when dealing with Neural Machine Translation (NMT) engines. This work-in-progress paper seeks to examine the English subtitles of the first episode of the popular Spanish Netflix series Cable Girls and the translated version generated by Google Translate and DeepL. Such analysis will help us determine whether there are significant linguistic differences that could lead to miscomprehension or cultural shocks. To this end, the corpus compiled consists of the Spanish script, the English subtitles available in Netflix and the translated version of the script. For the analysis of the data, errors have been classified following the DQF/MQM Error typology and have been evaluated with the automatic BLEU metric. Results show that NMT engines offer good-quality translations, which in turn may benefit translators working with audiovisual entertainment resources.
أُحدثت في العالم الإسلامي كليات للتربية و معاهد للتعليم؛ تبنت قيم الآخرين و تجاربهم، متجاهلة تراث الأمة التربوي. و التاريخ أصدق شاهد، و الواقع خير برهان، ففي حين عادى الغرب العلم، و أقصى العلماء، أرسى الإسلام قواعد منهج تربوي ثبتت نجاعته، و وضع أسساً للبحث العلمي، و هدفت هذه الدراسة إلى توضيح أهم معالم التربية النبوية في تفعيل الوسائل السمعية و البصرية.

suggested questions

comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا