Do you want to publish a course? Click here

Memory-Rate Tradeoff for Caching with Uncoded Placement under Nonuniform Random Demands

73   0   0.0 ( 0 )
 Added by Yong Deng
 Publication date 2021
and research's language is English




Ask ChatGPT about the research

For a caching system with multiple users, we aim to characterize the memory-rate tradeoff for caching with uncoded cache placement, under nonuniform file popularity. Focusing on the modified coded caching scheme (MCCS) recently proposed by Yu, etal., we formulate the cache placement optimization problem for the MCCS to minimize the average delivery rate under nonuniform file popularity, restricting to a class of popularity-first placements. We then present two information-theoretic lower bounds on the average rate for caching with uncoded placement, one for general cache placements and the other restricted to the popularity-first placements. By comparing the average rate of the optimized MCCS with the lower bounds, we prove that the optimized MCCS attains the general lower bound for the two-user case, providing the exact memory-rate tradeoff. Furthermore, it attains the popularity-first-based lower bound for the case of general K users with distinct file requests. In these two cases, our results also reveal that the popularity-first placement is optimal for the MCCS, and zero-padding used in coded delivery incurs no loss of optimality. For the case of K users with redundant file requests, our analysis shows that there may exist a gap between the optimized MCCS and the lower bounds due to zero-padding. We next fully characterize the optimal popularity-first cache placement for the MCCS, which is shown to possess a simple file-grouping structure and can be computed via an efficient algorithm using closed-form expressions. Finally, we extend our study to accommodate both nonuniform file popularity and sizes, where we show that the optimized MCCS attains the lower bound for the two-user case, providing the exact memory-rate tradeoff. Numerical results show that, for general settings, the gap between the optimized MCCS and the lower bound only exists in limited cases and is very small.



rate research

Read More

We consider a cache-aided wireless device-to-device (D2D) network under the constraint of one-shot delivery, where the placement phase is orchestrated by a central server. We assume that the devices caches are filled with uncoded data, and the whole file database at the server is made available in the collection of caches. Following this phase, the files requested by the users are serviced by inter-device multicast communication. For such a system setting, we provide the exact characterization of load-memory trade-off, by deriving both the minimum average and the minimum peak sum-loads of links between devices, for a given individual memory size at disposal of each user.
Recently Hachem et al. formulated a multiaccess coded caching model which consists of a central server connected to $K$ users via an error-free shared link, and $K$ cache-nodes. Each cache-node is equipped with a local cache and each user can access $L$ neighbouring cache-nodes with a cyclic wrap-around fashion. In this paper, we take the privacy of the users demands into consideration, i.e., each user can only get its required file and can not get any information about the demands of other users. By storing some private keys at the cache-nodes, we propose a novel transformation approach to transform a non-private multiaccess coded caching scheme into a private multiaccess coded caching scheme.
Caching at the wireless edge nodes is a promising way to boost the spatial and spectral efficiency, for the sake of alleviating networks from content-related traffic. Coded caching originally introduced by Maddah-Ali and Niesen significantly speeds up communication efficiency by transmitting multicast messages simultaneously useful to multiple users. Most prior works on coded caching are based on the assumption that each user may request all content in the library. However, in many applications the users are interested only in a limited set of content items that depends on their location. For example, visitors in a museum may stream audio and video related to the artworks in the room they are visiting, or assisted self-driving vehicles may access super-high definition maps of the area through which they are travelling. Motivated by these considerations, this paper formulates the coded caching problem for location-based content with edge cache nodes. The considered problem includes a content server with access to N location-based files, K edge cache nodes located at different regions, and K users each of which is in the serving region of one cache node and can retrieve the cached content of this cache node with negligible cost. Depending on the location, each user only requests a file from a location-dependent subset of the library. The objective is to minimize the worst-case load transmitted from the content server among all possible demands. We propose a highly non-trivial converse bound under uncoded cache placement, which shows that a simple achievable scheme is optimal. In addition, this achievable scheme is generally order optimal within 3. Finally, we extend the coded caching problem for location-based content to the multiaccess coded caching topology, where each user is connected to L nearest cache nodes. When $L geq 2$ we characterize the exact optimality on the worst-case load.
Coded caching schemes with low subpacketization and small transmission rate are desirable in practice due to the requirement of low implementation complexity and efficiency of the transmission. Placement delivery arrays (PDA in short) can be used to generate coded caching schemes. However, many known coded caching schemes have large memory ratios. In this paper, we realize that some schemes with low subpacketization generated by PDAs do not fully use the users caching content to create multicasting opportunities and thus propose to overcome this drawback. As an application, we obtain two new schemes with low subpacketizations, which have significantly advantages on the memory ratio and transmission rate compared with the original scheme.
Virtual reality (VR) over wireless is emerging as an important use case of 5G networks. Immersive VR experience requires the delivery of huge data at ultra-low latency, thus demanding ultra-high transmission rate. This challenge can be largely addressed by the recent network architecture known as mobile edge computing (MEC), which enables caching and computing capabilities at the edge of wireless networks. This paper presents a novel MEC-based mobile VR delivery framework that is able to cache parts of the field of views (FOVs) in advance and run certain post-processing procedures at the mobile VR device. To optimize resource allocation at the mobile VR device, we formulate a joint caching and computing decision problem to minimize the average required transmission rate while meeting a given latency constraint. When FOVs are homogeneous, we obtain a closed-form expression for the optimal joint policy which reveals interesting communications-caching-computing tradeoffs. When FOVs are heterogeneous, we obtain a local optima of the problem by transforming it into a linearly constrained indefinite quadratic problem then applying concave convex procedure. Numerical results demonstrate great promises of the proposed mobile VR delivery framework in saving communication bandwidth while meeting low latency requirement.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا