No Arabic abstract
It is a great challenge to evaluate the network performance of cellular mobile communication systems. In this paper, we propose new spatial spectrum and energy efficiency models for Poisson-Voronoi tessellation (PVT) random cellular networks. To evaluate the user access the network, a Markov chain based wireless channel access model is first proposed for PVT random cellular networks. On that basis, the outage probability and blocking probability of PVT random cellular networks are derived, which can be computed numerically. Furthermore, taking into account the call arrival rate, the path loss exponent and the base station (BS) density in random cellular networks, spatial spectrum and energy efficiency models are proposed and analyzed for PVT random cellular networks. Numerical simulations are conducted to evaluate the network spectrum and energy efficiency in PVT random cellular networks.
Wireless traffic attributable to machine learning (ML) inference workloads is increasing with the proliferation of applications and smart wireless devices leveraging ML inference. Owing to limited compute capabilities at these edge devices, achieving high inference accuracy often requires coordination with a remote compute node or cloud over the wireless cellular network. The accuracy of this distributed inference is, thus, impacted by the communication rate and reliability offered by the cellular network. In this paper, an analytical framework is proposed to characterize inference accuracy as a function of cellular network design. Using the developed framework, it is shown that cellular network should be provisioned with a minimum density of access points (APs) to guarantee a target inference accuracy, and the inference accuracy achievable at asymptotically high AP density is limited by the air-interface bandwidth. Furthermore, the minimum accuracy required of edge inference to deliver a target inference accuracy is shown to be inversely proportional to the density of APs and the bandwidth.
The growing demand for high-speed data, quality of service (QoS) assurance and energy efficiency has triggered the evolution of 4G LTE-A networks to 5G and beyond. Interference is still a major performance bottleneck. This paper studies the application of physical-layer network coding (PNC), a technique that exploits interference, in heterogeneous cellular networks. In particular, we propose a rate-maximising relay selection algorithm for a single cell with multiple relays based on the decode-and-forward strategy. With nodes transmitting at different powers, the proposed algorithm adapts the resource allocation according to the differing link rates and we prove theoretically that the optimisation problem is log-concave. The proposed technique is shown to perform significantly better than the widely studied selection-cooperation technique. We then undertake an experimental study on a software radio platform of the decoding performance of PNC with unbalanced SNRs in the multiple-access transmissions. This problem is inherent in cellular networks and it is shown that with channel coding and decoders based on multiuser detection and successive interference cancellation, the performance is better with power imbalance. This paper paves the way for further research in multi-cell PNC, resource allocation, and the implementation of PNC with higher-order modulations and advanced coding techniques.
Cellular vehicle-to-everything (C-V2X) communication, as a part of 5G wireless communication, has been considered one of the most significant techniques for Smart City. Vehicles platooning is an application of Smart City that improves traffic capacity and safety by C-V2X. However, different from vehicles platooning travelling on highways, C-V2X could be more easily eavesdropped and the spectrum resource could be limited when they converge at an intersection. Satisfying the secrecy rate of C-V2X, how to increase the spectrum efficiency (SE) and energy efficiency (EE) in the platooning network is a big challenge. In this paper, to solve this problem, we propose a Security-Aware Approach to Enhancing SE and EE Based on Deep Reinforcement Learning, named SEED. The SEED formulates an objective optimization function considering both SE and EE, and the secrecy rate of C-V2X is treated as a critical constraint of this function. The optimization problem is transformed into the spectrum and transmission power selections of V2V and V2I links using deep Q network (DQN). The heuristic result of SE and EE is obtained by the DQN policy based on rewards. Finally, we simulate the traffic and communication environments using Python. The evaluation results demonstrate that the SEED outperforms the DQN-wopa algorithm and the baseline algorithm by 31.83 % and 68.40 % in efficiency. Source code for the SEED is available at https://github.com/BandaidZ/OptimizationofSEandEEBasedonDRL.
The theory of wireless information and power transfer in energy constrained wireless networks has caught the interest of researchers due to its potential in increasing the lifetime of sensor nodes and mitigate the environment hazards caused by conventional cell batteries. Similarly, the advancements in areas of cooperative spectrum sharing protocols has enabled efficient use of frequency spectrum between a licensed primary user and a secondary user. In this paper, we consider an energy constrained secondary user which harvests energy from the primary signal and relays the primary signal in exchange for the spectrum access. We consider Nakagami-m fading model and propose two key protocols, namely time-splitting cooperative spectrum sharing (TS-CSS) and power-sharing cooperative spectrum sharing (PS-CSS), and derive expressions for the outage probabilities of the primary and secondary user in decode-forward and amplify-forward relaying modes. From the obtained results, it has been shown that the secondary user can carry its own transmission without adversely affecting the performance of the primary user and that PS-CSS protocol outperforms the TS-PSS protocol in terms of outage probability over a wide range of Signal to noise ratio(SNRs). The effect of various system parameters on the outage performance of these protocols have also been studied.
In this paper, we present a multi-user resource allocation framework using fragmented-spectrum synchronous OFDM-CDMA modulation over a frequency-selective fading channel. In particular, given pre-existing communications in the spectrum where the system is operating, a channel sensing and estimation method is used to obtain information of subcarrier availability. Given this information, some real-valued multi-level orthogonal codes, which are orthogonal codes with values of ${pm1,pm2,pm3,pm4, ... }$, are provided for emerging new users, i.e., cognitive radio users. Additionally, we have obtained a closed form expression for bit error rate of cognitive radio receivers in terms of detection probability of primary users, CR users sensing time and CR users signal to noise ratio. Moreover, simulation results obtained in this paper indicate the precision with which the analytical results have been obtained in modeling the aforementioned system.