ترغب بنشر مسار تعليمي؟ اضغط هنا

All Technologies Work Together for Good: A Glance to Future Mobile Networks

75   0   0.0 ( 0 )
 نشر من قبل Animesh Yadav
 تاريخ النشر 2018
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

The astounding capacity requirements of 5G have motivated researchers to investigate the feasibility of many potential technologies, such as massive multiple-input multiple-output, millimeter wave, full-duplex, non-orthogonal multiple access, carrier aggregation, cognitive radio, and network ultra-densification. The benefits and challenges of these technologies have been thoroughly studied either individually or in a combination of two or three. It is not clear, however, whether all potential technologies operating together lead to fulfilling the requirements posed by 5G. This paper explores the potential benefits and challenges when all technologies coexist in an ultra-dense cellular environment. The sum rate of the network is investigated with respect to the increase in the number of small-cells and results show the capacity gains achieved by the coexistence.



قيم البحث

اقرأ أيضاً

Interference between nodes directly limits the capacity of mobile ad hoc networks. This paper focuses on spatial interference cancelation with perfect channel state information (CSI), and analyzes the corresponding network capacity. Specifically, by using multiple antennas, zero-forcing beamforming is applied at each receiver for canceling the strongest interferers. Given spatial interference cancelation, the network transmission capacity is analyzed in this paper, which is defined as the maximum transmitting node density under constraints on outage and the signal-to-interference-noise ratio. Assuming the Poisson distribution for the locations of network nodes and spatially i.i.d. Rayleigh fading channels, mathematical tools from stochastic geometry are applied for deriving scaling laws for transmission capacity. Specifically, for small target outage probability, transmission capacity is proved to increase following a power law, where the exponent is the inverse of the size of antenna array or larger depending on the pass loss exponent. As shown by simulations, spatial interference cancelation increases transmission capacity by an order of magnitude or more even if only one extra antenna is added to each node.
The emergence of novel wireless networking paradigms such as small cell and cognitive radio networks has forever transformed the way in which wireless systems are operated. In particular, the need for self-organizing solutions to manage the scarce sp ectral resources has become a prevalent theme in many emerging wireless systems. In this paper, the first comprehensive tutorial on the use of matching theory, a Nobelprize winning framework, for resource management in wireless networks is developed. To cater for the unique features of emerging wireless networks, a novel, wireless-oriented classification of matching theory is proposed. Then, the key solution concepts and algorithmic implementations of this framework are exposed. Then, the developed concepts are applied in three important wireless networking areas in order to demonstrate the usefulness of this analytical tool. Results show how matching theory can effectively improve the performance of resource allocation in all three applications discussed.
175 - Hongjian Sun , Bo Tan , Jing Jiang 2013
Wireless technologies can support a broad range of smart grid applications including advanced metering infrastructure (AMI) and demand response (DR). However, there are many formidable challenges when wireless technologies are applied to the smart gi rd, e.g., the tradeoffs between wireless coverage and capacity, the high reliability requirement for communication, and limited spectral resources. Relaying has emerged as one of the most promising candidate solutions for addressing these issues. In this article, an introduction to various relaying strategies is presented, together with a discussion of how to improve spectral efficiency and coverage in relay-based information and communications technology (ICT) infrastructure for smart grid applications. Special attention is paid to the use of unidirectional relaying, collaborative beamforming, and bidirectional relaying strategies.
In forthcoming years, the Internet of Things (IoT) will connect billions of smart devices generating and uploading a deluge of data to the cloud. If successfully extracted, the knowledge buried in the data can significantly improve the quality of lif e and foster economic growth. However, a critical bottleneck for realising the efficient IoT is the pressure it puts on the existing communication infrastructures, requiring transfer of enormous data volumes. Aiming at addressing this problem, we propose a novel architecture dubbed Condense, which integrates the IoT-communication infrastructure into data analysis. This is achieved via the generic concept of network function computation: Instead of merely transferring data from the IoT sources to the cloud, the communication infrastructure should actively participate in the data analysis by carefully designed en-route processing. We define the Condense architecture, its basic layers, and the interactions among its constituent modules. Further, from the implementation side, we describe how Condense can be integrated into the 3rd Generation Partnership Project (3GPP) Machine Type Communications (MTC) architecture, as well as the prospects of making it a practically viable technology in a short time frame, relying on Network Function Virtualization (NFV) and Software Defined Networking (SDN). Finally, from the theoretical side, we survey the relevant literature on computing atomic functions in both analog and digital domains, as well as on function decomposition over networks, highlighting challenges, insights, and future directions for exploiting these techniques within practical 3GPP MTC architecture.
In network function computation is as a means to reduce the required communication flow in terms of number of bits transmitted per source symbol. However, the rate region for the function computation problem in general topologies is an open problem, and has only been considered under certain restrictive assumptions (e.g. tree networks, linear functions, etc.). In this paper, we propose a new perspective for distributing computation, and formulate a flow-based delay cost minimization problem that jointly captures the costs of communications and computation. We introduce the notion of entropic surjectivity as a measure to determine how sparse the function is and to understand the limits of computation. Exploiting Littles law for stationary systems, we provide a connection between this new notion and the computation processing factor that reflects the proportion of flow that requires communications. This connection gives us an understanding of how much a node (in isolation) should compute to communicate the desired function within the network without putting any assumptions on the topology. Our analysis characterizes the functions only via their entropic surjectivity, and provides insight into how to distribute computation. We numerically test our technique for search, MapReduce, and classification tasks, and infer for each task how sensitive the processing factor to the entropic surjectivity is.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا