ترغب بنشر مسار تعليمي؟ اضغط هنا

How to Distribute Computation in Networks

50   0   0.0 ( 0 )
 نشر من قبل Derya Malak
 تاريخ النشر 2019
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

In network function computation is as a means to reduce the required communication flow in terms of number of bits transmitted per source symbol. However, the rate region for the function computation problem in general topologies is an open problem, and has only been considered under certain restrictive assumptions (e.g. tree networks, linear functions, etc.). In this paper, we propose a new perspective for distributing computation, and formulate a flow-based delay cost minimization problem that jointly captures the costs of communications and computation. We introduce the notion of entropic surjectivity as a measure to determine how sparse the function is and to understand the limits of computation. Exploiting Littles law for stationary systems, we provide a connection between this new notion and the computation processing factor that reflects the proportion of flow that requires communications. This connection gives us an understanding of how much a node (in isolation) should compute to communicate the desired function within the network without putting any assumptions on the topology. Our analysis characterizes the functions only via their entropic surjectivity, and provides insight into how to distribute computation. We numerically test our technique for search, MapReduce, and classification tasks, and infer for each task how sensitive the processing factor to the entropic surjectivity is.



قيم البحث

اقرأ أيضاً

The astounding capacity requirements of 5G have motivated researchers to investigate the feasibility of many potential technologies, such as massive multiple-input multiple-output, millimeter wave, full-duplex, non-orthogonal multiple access, carrier aggregation, cognitive radio, and network ultra-densification. The benefits and challenges of these technologies have been thoroughly studied either individually or in a combination of two or three. It is not clear, however, whether all potential technologies operating together lead to fulfilling the requirements posed by 5G. This paper explores the potential benefits and challenges when all technologies coexist in an ultra-dense cellular environment. The sum rate of the network is investigated with respect to the increase in the number of small-cells and results show the capacity gains achieved by the coexistence.
187 - Roy D. Yates 2021
A source node updates its status as a point process and also forwards its updates to a network of observer nodes. Within the network of observers, these updates are forwarded as point processes from node to node. Each node wishes its knowledge of the source to be as timely as possible. In this network, timeliness is measured by a discrete form of age of information: each status change at the source is referred to as a version and the age at a node is how ma
Millimeter-wave (mmWave) networks rely on directional transmissions, in both control plane and data plane, to overcome severe path-loss. Nevertheless, the use of narrow beams complicates the initial cell-search procedure where we lack sufficient info rmation for beamforming. In this paper, we investigate the feasibility of random beamforming for cell-search. We develop a stochastic geometry framework to analyze the performance in terms of failure probability and expected latency of cell-search. Meanwhile, we compare our results with the naive, but heavily used, exhaustive search scheme. Numerical results show that, for a given discovery failure probability, random beamforming can substantially reduce the latency of exhaustive search, especially in dense networks. Our work demonstrates that developing complex cell-discovery algorithms may be unnecessary in dense mmWave networks and thus shed new lights on mmWave system design.
In this paper we investigate the performance of caching schemes based on fountain codes in a heterogeneous satellite network. We consider multiple cache-aided hubs which are connected to a geostationary satellite through backhaul links. With the aimo f reducing the average number of transmissions over the satellite backhaul link, we propose the use of a caching scheme based on fountain codes. We derive a simple analytical expression of the average backhaul transmission rate and provide a tightupper bound on it. Furthermore, we show how the performance of the fountain code based caching scheme is similar to that of a caching scheme based on maximum distance separable codes.
Timeliness is an emerging requirement for many Internet of Things (IoT) applications. In IoT networks, where a large-number of nodes are distributed, severe interference may incur during the transmission phase which causes age of information (AoI) de gradation. It is therefore important to study the performance limit of AoI as well as how to achieve such limit. In this paper, we aim to optimize the AoI in random access Poisson networks. By taking into account the spatio-temporal interactions amongst the transmitters, an expression of the peak AoI is derived, based on explicit expressions of the optimal peak AoI and the corresponding optimal system parameters including the packet arrival rate and the channel access probability are further derived. It is shown that with a given packet arrival rate (resp. a given channel access probability), the optimal channel access probability (resp. the optimal packet arrival rate), is equal to one under a small node deployment density, and decrease monotonically as the spatial deployment density increases due to the severe interference caused by spatio-temproal coupling between transmitters. When joint tuning of the packet arrival rate and channel access probability is performed, the optimal channel access probability is always set to be one. Moreover, with the sole tuning of the channel access probability, it is found that the optimal peak AoI performance can be improved with a smaller packet arrival rate only when the node deployment density is high, which is contrast to the case of the sole tuning of the packet arrival rate, where a higher channel access probability always leads to better optimal peak AoI regardless of the node deployment density. In all the cases of optimal tuning of system parameters, the optimal peak AoI linearly grows with the node deployment density as opposed to an exponential growth with fixed system parameters.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا