No Arabic abstract
In this paper, we consider a cache-aided relay network, where a single server consisting of a library of N files connects with K1 relays through a shared noiseless link, and each relay connects with K2 users through a shared noiseless link. Each relay and user are equipped with a cache memory of M1 and M2 files, respectively. We propose a centralized and a decentralized coded caching scheme that exploit the spared transmission time resource by allowing concurrent transmission between the two layers. It is shown that both caching schemes are approximately optimal, and greatly reduce the transmission delay compared to the previously known caching schemes. Surprisingly, we show that when the relays caching size is equal to a threshold that is strictly smaller than N (e.g. M1=0.382N under the decentralized setup and (K1-1)N/K1 under the centralized setup, when K1=2), our schemes achieve the same delay as if each relay had access to the full library. To our best knowledge, this is the first result showing that even the caching size is strictly smaller than the librarys size, increasing the caching size is wasteful in reducing the transmission latency.
We study downlink beamforming in a single-cell network with a multi-antenna base station (BS) serving cache-enabled users. For a given common rate of the files in the system, we first formulate the minimum transmit power with beamforming at the BS as a non-convex optimization problem. This corresponds to a multiple multicast problem, to which a stationary solution can be efficiently obtained through successive convex approximation (SCA). It is observed that the complexity of the problem grows exponentially with the number of subfiles delivered to each user in each time slot, which itself grows exponentially with the number of users in the system. Therefore, we introduce a low-complexity alternative through time-sharing that limits the number of subfiles that can be received by a user in each time slot. It is shown through numerical simulations that, the reduced-complexity beamforming scheme has minimal performance gap compared to transmitting all the subfiles jointly, and outperforms the state-of-the-art low-complexity scheme at all SNR and rate values with sufficient spatial degrees of freedom, and in the high SNR/high rate regime when the number of spatial degrees of freedom is limited.
In this paper, we investigate a large intelligent surface-enhanced (LIS-enhanced) system, where a LIS is deployed to assist secure transmission. Our design aims to maximize the achievable secrecy rates in different channel models, i.e., Rician fading and (or) independent and identically distributed Gaussian fading for the legitimate and eavesdropper channels. In addition, we take into consideration an artificial noise-aided transmission structure for further improving system performance. The difficulties of tackling the aforementioned problems are the structure of the expected secrecy rate expressions and the non-convex phase shift constraint. To facilitate the design, we propose two frameworks, namely the sample average approximation based (SAA-based) algorithm and the hybrid stochastic projected gradient-convergent policy (hybrid SPG-CP) algorithm, to calculate the expectation terms in the secrecy rate expressions. Meanwhile, majorization minimization (MM) is adopted to address the non-convexity of the phase shift constraint. In addition, we give some analyses on two special scenarios by making full use of the expectation terms. Simulation results show that the proposed algorithms effectively optimize the secrecy communication rate for the considered setup, and the LIS-enhanced system greatly improves secrecy performance compared to conventional architectures without LIS.
Rate-splitting multiple access (RSMA) has been recognized as a promising physical layer strategy for 6G. Motivated by ever increasing popularity of cache-enabled content delivery in wireless communications, this paper proposes an innovative multigroup multicast transmission scheme based on RSMA for cache-aided cloud-radio access networks (C-RAN). Our proposed scheme not only exploits the properties of content-centric communications and local caching at the base stations (BSs), but also incorporates RSMA to better manage interference in multigroup multicast transmission with statistical channel state information (CSI) known at the central processor (CP) and the BSs. At the RSMA-enabled cloud CP, the message of each multicast group is split into a private and a common part with the former private part being decoded by all users in the respective group and the latter common part being decoded by multiple users from other multicast groups. Common message decoding is done for the purpose of mitigating the interference. In this work, we jointly optimize the clustering of BSs and the precoding with the aim of maximizing the minimum rate among all multicast groups to guarantee fairness serving all groups. The problem is a mixed-integer non-linear stochastic program (MINLSP), which is solved by a practical algorithm we proposed including a heuristic clustering algorithm for assigning a set of BSs to serve each user followed by an efficient iterative algorithm that combines the sample average approximation (SAA) and weighted minimum mean square error (WMMSE) to solve the stochastic non-convex sub-problem of precoder design. Numerical results show the explicit max-min rate gain of our proposed transmission scheme compared to the state-of-the-art trivial interference processing methods. Therefore, we conclude that RSMA is a promising technique for cache-aided C-RAN.
This work investigates a system where each user aims to retrieve a scalar linear function of the files of a library, which are Maximum Distance Separable coded and stored at multiple distributed servers. The system needs to guarantee robust decoding in the sense that each user must decode its demanded function with signals received from any subset of servers whose cardinality exceeds a threshold. In addition, (a) the content of the library must be kept secure from a wiretapper who obtains all the signals from the servers;(b) any subset of users together can not obtain any information about the demands of the remaining users; and (c) the users demands must be kept private against all the servers even if they collude. Achievable schemes are derived by modifying existing Placement Delivery Array (PDA) constructions, originally proposed for single-server single-file retrieval coded caching systems without any privacy or security or robustness constraints. It is shown that the PDAs describing the original Maddah-Ali and Niesens coded caching scheme result in a load-memory tradeoff that is optimal to within a constant multiplicative gap, except for the small memory regime when the number of file is smaller than the number of users. As by-products, improved order optimality results are derived for three less restrictive systems in all parameter regimes.
This work investigates the problem of cache-aided content Secure and demand Private Linear Function Retrieval (SP-LFR), where three constraints are imposed on the system:(a) each user is interested in retrieving an arbitrary linear combination of the files in the servers library;(b) the content of the library must be kept secure from a wiretapper who obtains the signal sent by the server; and (c) no colluding subset of users together obtain information about the demands of the remaining users. A procedure is proposed to derive an SP-LFR scheme from a given Placement Delivery Array (PDA), which is known to give coded caching schemes with low subpacketization for systems with neither security nor privacy constraints. This procedure uses the superposition of security keys and privacy keys in both the cache placement and transmitted signal to guarantee content security and demand privacy, respectively. In particular, among all PDA-based SP-LFR schemes, the memory-load pairs achieved by the PDA describing the Maddah-Ali and Niesens scheme are Pareto-optimal and have the lowest subpacketization. Moreover, the achieved load-memory tradeoff is optimal to within a constant multiplicative gap except for the small memory regime (i.e., when the cache size is between 1 and 2) and the number of files is smaller than the number of users. Remarkably, the memory-load tradeoff does not increase compared to the best known schemes that guarantee either only content security in all regimes or only demand privacy in regime mentioned above.