ترغب بنشر مسار تعليمي؟ اضغط هنا

Federated edge learning (FEEL) is a widely adopted framework for training an artificial intelligence (AI) model distributively at edge devices to leverage their data while preserving their data privacy. The execution of a power-hungry learning task a t energy-constrained devices is a key challenge confronting the implementation of FEEL. To tackle the challenge, we propose the solution of powering devices using wireless power transfer (WPT). To derive guidelines on deploying the resultant wirelessly powered FEEL (WP-FEEL) system, this work aims at the derivation of the tradeoff between the model convergence and the settings of power sources in two scenarios: 1) the transmission power and density of power-beacons (dedicated charging stations) if they are deployed, or otherwise 2) the transmission power of a server (access-point). The development of the proposed analytical framework relates the accuracy of distributed stochastic gradient estimation to the WPT settings, the randomness in both communication and WPT links, and devices computation capacities. Furthermore, the local-computation at devices (i.e., mini-batch size and processor clock frequency) is optimized to efficiently use the harvested energy for gradient estimation. The resultant learning-WPT tradeoffs reveal the simple scaling laws of the model-convergence rate with respect to the transferred energy as well as the devices computational energy efficiencies. The results provide useful guidelines on WPT provisioning to provide a guaranteer on learning performance. They are corroborated by experimental results using a real dataset.
Wireless power transfer (WPT) is an emerging paradigm that will enable using wireless to its full potential in future networks, not only to convey information but also to deliver energy. Such networks will enable trillions of future low-power devices to sense, compute, connect, and energize anywhere, anytime, and on the move. The design of such future networks brings new challenges and opportunities for signal processing, machine learning, sensing, and computing so as to make the best use of the RF radiations, spectrum, and network infrastructure in providing cost-effective and real-time power supplies to wireless devices and enable wireless-powered applications. In this paper, we first review recent signal processing techniques to make WPT and wireless information and power transfer as efficient as possible. Topics include power amplifier and energy harvester nonlinearities, active and passive beamforming, intelligent reflecting surfaces, receive combining with multi-antenna harvester, modulation, coding, waveform, massive MIMO, channel acquisition, transmit diversity, multi-user power region characterization, coordinated multipoint, and distributed antenna systems. Then, we overview two different design methodologies: the model and optimize approach relying on analytical system models, modern convex optimization, and communication theory, and the learning approach based on data-driven end-to-end learning and physics-based learning. We discuss the pros and cons of each approach, especially when accounting for various nonlinearities in wireless-powered networks, and identify interesting emerging opportunities for the approaches to complement each other. Finally, we identify new emerging wireless technologies where WPT may play a key role -- wireless-powered mobile edge computing and wireless-powered sensing -- arguing WPT, communication, computation, and sensing must be jointly designed.
Edge machine learning involves the deployment of learning algorithms at the network edge to leverage massive distributed data and computation resources to train artificial intelligence (AI) models. Among others, the framework of federated edge learni ng (FEEL) is popular for its data-privacy preservation. FEEL coordinates global model training at an edge server and local model training at edge devices that are connected by wireless links. This work contributes to the energy-efficient implementation of FEEL in wireless networks by designing joint computation-and-communication resource management ($text{C}^2$RM). The design targets the state-of-the-art heterogeneous mobile architecture where parallel computing using both a CPU and a GPU, called heterogeneous computing, can significantly improve both the performance and energy efficiency. To minimize the sum energy consumption of devices, we propose a novel $text{C}^2$RM framework featuring multi-dimensional control including bandwidth allocation, CPU-GPU workload partitioning and speed scaling at each device, and $text{C}^2$ time division for each link. The key component of the framework is a set of equilibriums in energy rates with respect to different control variables that are proved to exist among devices or between processing units at each device. The results are applied to designing efficient algorithms for computing the optimal $text{C}^2$RM policies faster than the standard optimization tools. Based on the equilibriums, we further design energy-efficient schemes for device scheduling and greedy spectrum sharing that scavenges spectrum holes resulting from heterogeneous $text{C}^2$ time divisions among devices. Using a real dataset, experiments are conducted to demonstrate the effectiveness of $text{C}^2$RM on improving the energy efficiency of a FEEL system.
Edge machine learning involves the deployment of learning algorithms at the wireless network edge so as to leverage massive mobile data for enabling intelligent applications. The mainstream edge learning approach, federated learning, has been develop ed based on distributed gradient descent. Based on the approach, stochastic gradients are computed at edge devices and then transmitted to an edge server for updating a global AI model. Since each stochastic gradient is typically high-dimensional (with millions to billions of coefficients), communication overhead becomes a bottleneck for edge learning. To address this issue, we propose in this work a novel framework of hierarchical stochastic gradient quantization and study its effect on the learning performance. First, the framework features a practical hierarchical architecture for decomposing the stochastic gradient into its norm and normalized block gradients, and efficiently quantizes them using a uniform quantizer and a low-dimensional codebook on a Grassmann manifold, respectively. Subsequently, the quantized normalized block gradients are scaled and cascaded to yield the quantized normalized stochastic gradient using a so-called hinge vector designed under the criterion of minimum distortion. The hinge vector is also efficiently compressed using another low-dimensional Grassmannian quantizer. The other feature of the framework is a bit-allocation scheme for reducing the quantization error. The scheme determines the resolutions of the low-dimensional quantizers in the proposed framework. The framework is proved to guarantee model convergency by analyzing the convergence rate as a function of the quantization bits. Furthermore, by simulation, our design is shown to substantially reduce the communication overhead compared with the state-of-the-art signSGD scheme, while both achieve similar learning accuracies.
As the ratification of 5G New Radio technology is being completed, enabling network architectures are expected to undertake a matching effort. Conventional cloud and edge computing paradigms may thus become insufficient in supporting the increasingly stringent operating requirements of emph{intelligent~Internet-of-Things (IoT) devices} that can move unpredictably and at high speeds. Complementing these, the concept of fog emerges to deploy cooperative cloud-like functions in the immediate vicinity of various moving devices, such as connected and autonomous vehicles, on the road and in the air. Envisioning gradual evolution of these infrastructures toward the increasingly denser geographical distribution of fog functionality, we in this work put forward the vision of dense moving fog for intelligent IoT applications. To this aim, we review the recent powerful enablers, outline the main challenges and opportunities, and corroborate the performance benefits of collaborative dense fog operation in a characteristic use case featuring a connected fleet of autonomous vehicles.
66 - Yuqing Du , Kaibin Huang 2018
By implementing machine learning at the network edge, edge learning trains models by leveraging rich data distributed at edge devices (e.g., smartphones and sensors) and in return endow on them capabilities of seeing, listening, and reasoning. In edg e learning, the need of high-mobility wireless data acquisition arises in scenarios where edge devices (or even servers) are mounted on the ground or aerial vehicles. In this paper, we present a novel solution, called fast analog transmission (FAT), for high- mobility data acquisition in edge-learning systems, which has several key features. First, FAT incurs low-latency. Specifically, FAT requires no source-and-channel coding and no channel training via the proposed technique of Grassmann analog encoding (GAE) that encodes data samples into subspace matrices. Second, FAT supports spatial multiplexing by directly transmitting analog vector data over an antenna array. Third, FAT can be seamlessly integrated with edge learning (i.e., training of a classifier model in this work). In particular, by applying a Grassmannian-classification algorithm from computer vision, the received GAE encoded data can be directly applied to training the model without decoding and conversion. This design is found by simulation to outperform conventional schemes in learning accuracy due to its robustness against data distortion induced by fast fading.
As the realization of vehicular communication such as vehicle-to-vehicle (V2V) or vehicle-to-infrastructure (V2I) is imperative for the autonomous driving cars, the understanding of realistic vehicle-to-everything (V2X) models is needed. While previo us research has mostly targeted vehicular models in which vehicles are randomly distributed and the variable of carrier frequency was not considered, a more realistic analysis of the V2X model is proposed in this paper. We use a one-dimensional (1D) Poisson cluster process (PCP) to model a realistic scenario of vehicle distribution in a perpendicular cross line road urban area and compare the coverage results with the previous research that distributed vehicles randomly by Poisson Point Process (PPP). Moreover, we incorporate the effect of different carrier frequencies, mmWave and sub-6 GHz, to our analysis by altering the antenna radiation pattern accordingly. Results indicated that while the effect of clustering led to lower outage, using mmWave had even more significance in leading to lower outage. Moreover, line-of-sight (LoS) interference links are shown to be more dominant in lowering the outage than the non-line-of-sight (NLoS) links even though they are less in number. The analytical results give insight into designing and analyzing the urban V2X channels, and are verified by actual urban area three-dimensional (3D) ray-tracing simulation.
314 - Dongzhu Liu , Kaibin Huang 2017
Multimedia content especially videos is expected to dominate data traffic in next-generation mobile networks. Caching popular content at the network edge has emerged to be a solution for low-latency content delivery. Compared with the traditional wir eless communication, content delivery has a key characteristic that many signals coexisting in the air carry identical popular content. They, however, can interfere with each other at a receiver if their modulation-and-coding (MAC) schemes are adapted to individual channels following the classic approach. To address this issue, we present a novel idea of content adaptive MAC (CAMAC) where adapting MAC schemes to content ensures that all signals carry identical content are encoded using an identical MAC scheme, achieving spatial MAC alignment. Consequently, interference can be harnessed as signals, to improve the reliability of wireless delivery. In the remaining part of the paper, we focus on quantifying the gain CAMAC can bring to a content-delivery network using a stochastic-geometry model. Specifically, content helpers are distributed as a Poisson point process, each of which transmits a file from a content database based on a given popularity distribution. It is discovered that the successful content-delivery probability is closely related to the distribution of the ratio of two independent shot noise processes, named a shot-noise ratio. The distribution itself is an open mathematical problem that we tackle in this work. Using stable-distribution theory and tools from stochastic geometry, the distribution function is derived in closed form. Extending the result in the context of content-delivery networks with CAMAC yields the content-delivery probability in different closed forms. In addition, the gain in the probability due to CAMAC is shown to grow with the level of skewness in the content popularity distribution.
In this article, we put forward the mobile crowd sensing paradigm based on ubiquitous wearable devices carried by human users. The key challenge for mass user involvement into prospective urban crowd sending applications, such as monitoring of large- scale phenomena (e.g., traffic congestion and air pollution levels), is the appropriate sources of motivation. We thus advocate for the use of wireless power transfer provided in exchange for sensed data to incentivize the owners of wearables to participate in collaborative data collection. Based on this construction, we develop the novel concept of wirelessly powered crowd sensing and offer the corresponding network architecture considerations together with a systematic review of wireless charging techniques to implement it. Further, we contribute a detailed system-level feasibility study that reports on the achievable performance levels for the envisioned setup. Finally, the underlying energy-data trading mechanisms are discussed, and the work is concluded with outlining open research opportunities.
The vision of seamlessly integrating information transfer (IT) and microwave based power transfer (PT) in the same system has led to the emergence of a new research area, called wirelessly power communications (WPC). Extensive research has been condu cted on developing WPC theory and techniques, building on the extremely rich wireless communications litera- ture covering diversified topics such as transmissions, resource allocations, medium access control and network protocols and architectures. Despite these research efforts, transforming WPC from theory to practice still faces many unsolved prob- lems concerning issues such as mobile complexity, power transfer efficiency, and safety. Furthermore, the fundamental limits of WPC remain largely unknown. Recent attempts to address these open issues has resulted in the emergence of numerous new research trends in the WPC area. A few promising trends are introduced in this article. From the practical perspective, the use of backscatter antennas can support WPC for low-complexity passive devices, the design of spiky waveforms can improve the PT efficiency, and analog spatial decoupling is proposed for solving the PT-IT near-far problem in WPC. From the theoretic perspective, the fundamental limits of WPC can be quantified by leveraging recent results on super-directivity and the limit can be improved by the deployment of large-scale distributed antenna arrays. Specific research problems along these trends are discussed, whose solutions can lead to significant advancements in WPC.
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا