ترغب بنشر مسار تعليمي؟ اضغط هنا

Transparent caching of virtual stubs for improved performance in ubiquitous environments

35   0   0.0 ( 0 )
 نشر من قبل Lachhman Das Dhomeja
 تاريخ النشر 2011
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

Context-awareness is an essential requirement for pervasive computing applications, which enables them to adapt and perform tasks based on context. One of the adaptive features of context-awareness is contextual reconfiguration. Contextual reconfiguration involves discovering remote service(s) based on context and binding them to the application components to realize new behaviors, which may be needed to satisfy user needs or to enrich user experience. One of the steps in the reconfiguration process involves a remote lookup to discover the service(s) based on context. This remote lookup process provides the largest contribution to reconfiguration time and this is due to fact that the remote calls are much slower than local calls. Consequently, it affects system performance. In pervasive computing applications, this may turn out to be undesirable in terms of user experience. Moreover, other distributed applications using the network may be affected as every remote method call decreases the amount of bandwidth available on the network. Various systems provide reconfiguration support and offer high-level reconfiguration directives to develop adaptive context-aware applications, but do not address this performance bottleneck. We address this issue and implement seamless caching of virtual stubs within our PCRA1 for improved performance. In this paper we present and describe our transparent caching support and also provide its performance evaluation.

قيم البحث

اقرأ أيضاً

Synthetic data generation has become essential in last years for feeding data-driven algorithms, which surpassed traditional techniques performance in almost every computer vision problem. Gathering and labelling the amount of data needed for these d ata-hungry models in the real world may become unfeasible and error-prone, while synthetic data give us the possibility of generating huge amounts of data with pixel-perfect annotations. However, most synthetic datasets lack from enough realism in their rendered images. In that context UnrealROX generation tool was presented in 2019, allowing to generate highly realistic data, at high resolutions and framerates, with an efficient pipeline based on Unreal Engine, a cutting-edge videogame engine. UnrealROX enabled robotic vision researchers to generate realistic and visually plausible data with full ground truth for a wide variety of problems such as class and instance semantic segmentation, object detection, depth estimation, visual grasping, and navigation. Nevertheless, its workflow was very tied to generate image sequences from a robotic on-board camera, making hard to generate data for other purposes. In this work, we present UnrealROX+, an improved version of UnrealROX where its decoupled and easy-to-use data acquisition system allows to quickly design and generate data in a much more flexible and customizable way. Moreover, it is packaged as an Unreal plug-in, which makes it more comfortable to use with already existing Unreal projects, and it also includes new features such as generating albedo or a Python API for interacting with the virtual environment from Deep Learning frameworks.
Computational micromagnetics requires numerical solution of partial differential equations to resolve complex interactions in magnetic nanomaterials. The Virtual Micromagnetics project described here provides virtual machine simulation environments t o run open-source micromagnetic simulation packages. These environments allow easy access to simulation packages that are often difficult to compile and install, and enable simulations and their data to be shared and stored in a single virtual hard disk file, which encourages reproducible research. Virtual Micromagnetics can be extended to automate the installation of micromagnetic simulation packages on non-virtual machines, and to support closed-source and new open-source simulation packages, including packages from disciplines other than micromagnetics, encouraging reuse. Virtual Micromagnetics is stored in a public GitHub repository under a three-clause Berkeley Software Distribution (BSD) license.
85 - Falk Herwig 2018
Collaborations in astronomy and astrophysics are faced with numerous cyber infrastructure challenges, such as large data sets, the need to combine heterogeneous data sets, and the challenge to effectively collaborate on those large, heterogeneous dat a sets with significant processing requirements and complex science software tools. The cyberhubs system is an easy-to-deploy package for small to medium-sized collaborations based on the Jupyter and Docker technology, that allows web-browser enabled, remote, interactive analytic access to shared data. It offers an initial step to address these challenges. The features and deployment steps of the system are described, as well as the requirements collection through an account of the different approaches to data structuring, handling and available analytic tools for the NuGrid and PPMstar collaborations. NuGrid is an international collaboration that creates stellar evolution and explosion physics and nucleosynthesis simulation data. The PPMstar collaboration performs large-scale 3D stellar hydrodynamics simulation of interior convection in the late phases of stellar evolution. Examples of science that is presently performed on cyberhubs, in the areas 3D stellar hydrodynamic simulations, stellar evolution and nucleosynthesis and Galactic chemical evolution, are presented.
In an $(H,r)$ combination network, a single content library is delivered to ${Hchoose r}$ users through deployed $H$ relays without cache memories, such that each user with local cache memories is simultaneously served by a different subset of $r$ re lays on orthogonal non-interfering and error-free channels. The combinatorial placement delivery array (CPDA in short) can be used to realize a coded caching scheme for combination networks. In this paper, a new algorithm realizing a coded caching scheme for combination network based on a CPDA is proposed such that the schemes obtained have smaller subpacketization levels or are implemented more flexible than the previously known schemes. Then we focus on directly constructing CPDAs for any positive integers $H$ and $r$ with $r<H$. This is different from the grouping method in reference (IEEE ISIT, 17-22, 2018) under the constraint that $r$ divides $H$. Consequently two classes of CPDAs are obtained. Finally comparing to the schemes and the method proposed by Yan et al., (IEEE ISIT, 17-22, 2018) the schemes realized by our CPDAs have significantly advantages on the subpacketization levels and the transmission rates.
Improved lower bounds on the average and the worst-case rate-memory tradeoffs for the Maddah-Ali&Niesen coded caching scenario are presented. For any number of users and files and for arbitrary cache sizes, the multiplicative gap between the exact ra te-memory tradeoff and the new lower bound is less than 2.315 in the worst-case scenario and less than 2.507 in the average-case scenario.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا