No Arabic abstract
NarrowBand-Internet of Things (NB-IoT) is a new 3GPP radio access technology designed to provide better coverage for Low Power Wide Area (LPWA) networks. To provide reliable connections with extended coverage, a repetition transmission scheme and up to three Coverage Enhancement (CE) groups are introduced into NB-IoT during both Random Access CHannel (RACH) procedure and data transmission procedure, where each CE group is configured with different repetition values and transmission resources. To characterize the RACH performance of the NB-IoT network with three CE groups, this paper develops a novel traffic-aware spatio-temporal model to analyze the RACH success probability, where both the preamble transmission outage and the collision events of each CE group jointly determine the traffic evolution and the RACH success probability. Based on this analytical model, we derive the analytical expression for the RACH success probability of a randomly chosen IoT device in each CE group over multiple time slots with different RACH schemes, including baseline, back-off (BO), access class barring (ACB), and hybrid ACB and BO schemes (ACB&BO). Our results have shown that the RACH success probabilities of the devices in three CE groups outperform that of a single CE group network but not for all the groups, which is affected by the choice of the categorizing parameters.This mathematical model and analytical framework can be applied to evaluate the performance of multiple group users of other networks with spatial separations.
Wireless cellular networks have many parameters that are normally tuned upon deployment and re-tuned as the network changes. Many operational parameters affect reference signal received power (RSRP), reference signal received quality (RSRQ), signal-to-interference-plus-noise-ratio (SINR), and, ultimately, throughput. In this paper, we develop and compare two approaches for maximizing coverage and minimizing interference by jointly optimizing the transmit power and downtilt (elevation tilt) settings across sectors. To evaluate different parameter configurations offline, we construct a realistic simulation model that captures geographic correlations. Using this model, we evaluate two optimization methods: deep deterministic policy gradient (DDPG), a reinforcement learning (RL) algorithm, and multi-objective Bayesian optimization (BO). Our simulations show that both approaches significantly outperform random search and converge to comparable Pareto frontiers, but that BO converges with two orders of magnitude fewer evaluations than DDPG. Our results suggest that data-driven techniques can effectively self-optimize coverage and capacity in cellular networks.
The introduction of Narrowband Internet of Things (NB-IoT) as a cellular IoT technology aims to support massive Machine-Type Communications applications. These applications are characterized by massive connections from a large number of low-complexity and low-power devices. One of the goals of NB-IoT is to improve coverage extension beyond existing cellular technologies. In order to do that, NB-IoT introduces transmission repetitions and different bandwidth allocation configurations in uplink. These new transmission approaches yield many transmission options in uplink. In this paper, we propose analytical expressions that describe the influence of these new approaches in the transmission. Our analysis is based on the Shannon theorem. The transmission is studied in terms of the required Signal to Noise Ratio, bandwidth utilization, and energy per transmitted bit. Additionally, we propose an uplink link adaptation algorithm that contemplates these new transmission approaches. The conducted evaluation summarizes the influence of these approaches. Furthermore, we present the resulting uplink link adaptation from our proposed algorithm sweeping the devices coverage.
This paper introduces a general approach to design a tailored solution to detect rare events in different industrial applications based on Internet of Things (IoT) networks and machine learning algorithms. We propose a general framework based on three layers (physical, data and decision) that defines the possible designing options so that the rare events/anomalies can be detected ultra-reliably. This general framework is then applied in a well-known benchmark scenario, namely Tennessee Eastman Process. We then analyze this benchmark under three threads related to data processes: acquisition, fusion and analytics. Our numerical results indicate that: (i) event-driven data acquisition can significantly decrease the number of samples while filtering measurement noise, (ii) mutual information data fusion method can significantly decrease the variable spaces and (iii) quantitative association rule mining method for data analytics is effective for the rare event detection, identification and diagnosis. These results indicates the benefits of an integrated solution that jointly considers the different levels of data processing following the proposed general three layer framework, including details of the communication network and computing platform to be employed.
In future drone applications fast moving unmanned aerial vehicles (UAVs) will need to be connected via a high throughput ultra reliable wireless link. MmWave communication is assumed to be a promising technology for UAV communication, as the narrow beams cause little interference to and from the ground. A challenge for such networks is the beamforming requirement, and the fact that frequent handovers are required as the cells are small. In the UAV communication research community, mobility and especially handovers are often neglected, however when considering beamforming, antenna array sizes start to matter and the effect of azimuth and elevation should be studied, especially their impact on handover rate and outage capacity. This paper aims to fill some of this knowledge gap and to shed some light on the existing problems. This work will analyse the performance of 3D beamforming and handovers for UAV networks through a case study of a realistic 5G deployment using mmWave. We will look at the performance of a UAV flying over a city utilizing a beamformed mmWave link.
NarrowBand-Internet of Things (NB-IoT) is a new 3GPP radio access technology designed to provide better coverage for a massive number of low-throughput low-cost devices in delay-tolerant applications with low power consumption. To provide reliable connections with extended coverage, a repetition transmission scheme is introduced to NB-IoT during both Random Access CHannel (RACH) procedure and data transmission procedure. To avoid the difficulty in replacing the battery for IoT devices, the energy harvesting is considered as a promising solution to support energy sustainability in the NB-IoT network. In this work, we analyze RACH success probability in a self-powered NB-IoT network taking into account the repeated preamble transmissions and collisions, where each IoT device with data is active when its battery energy is sufficient to support the transmission. We model the temporal dynamics of the energy level as a birth-death process, derive the energy availability of each IoT device, and examine its dependence on the energy storage capacity and the repetition value. We show that in certain scenarios, the energy availability remains unchanged despite randomness in the energy harvesting. We also derive the exact expression for the RACH success probability of a {randomly chosen} IoT device under the derived energy availability, which is validated under different repetition values via simulations. We show that the repetition scheme can efficiently improve the RACH success probability in a light traffic scenario, but only slightly improves that performance with very inefficient channel resource utilization in a heavy traffic scenario.