No Arabic abstract
The proliferation of Content Delivery Networks (CDN) reveals that existing content networks are owned and operated by individual companies. As a consequence, closed delivery networks are evolved which do not cooperate with other CDNs and in practice, islands of CDNs are formed. Moreover, the logical separation between contents and services in this context results in two content networking domains. But present trends in content networks and content networking capabilities give rise to the interest in interconnecting content networks. Finding ways for distinct content networks to coordinate and cooperate with other content networks is necessary for better overall service. In addition to that, meeting the QoS requirements of users according to the negotiated Service Level Agreements between the user and the content network is a burning issue in this perspective. In this article, we present an open, scalable and Service-Oriented Architecture based system to assist the creation of open Content and Service Delivery Networks (CSDN) that scale and support sharing of resources with other CSDNs.
We address the problem of content replication in large distributed content delivery networks, composed of a data center assisted by many small servers with limited capabilities and located at the edge of the network. The objective is to optimize the placement of contents on the servers to offload as much as possible the data center. We model the system constituted by the small servers as a loss network, each loss corresponding to a request to the data center. Based on large system / storage behavior, we obtain an asymptotic formula for the optimal replication of contents and propose adaptive schemes related to those encountered in cache networks but reacting here to loss events, and faster algorithms generating virtual events at higher rate while keeping the same target replication. We show through simulations that our adaptive schemes outperform significantly standard replication strategies both in terms of loss rates and adaptation speed.
Blockchain is a merging technology for decentralized management and data security, which was first introduced as the core technology of cryptocurrency, e.g., Bitcoin. Since the first success in financial sector, blockchain has shown great potentials in various domains, e.g., internet of things and mobile networks. In this paper, we propose a novel blockchain-based architecture for content delivery networks (B-CDN), which exploits the advances of the blockchain technology to provide a decentralized and secure platform to connect content providers (CPs) with users. On one hand, the proposed B-CDN will leverage the registration and subscription of the users to different CPs, while guaranteeing the user privacy thanks to virtual identity provided by the blockchain network. On the other hand, the B-CDN creates a public immutable database of the requested contents (from all CPs), based on which each CP can better evaluate the user preference on its contents. The benefits of B-CDN are demonstrated via an edge-caching application, in which a feature-based caching algorithm is proposed for all CPs. The proposed caching algorithm is verified with the realistic Movielens dataset. A win-win relation between the CPs and users is observed, where the B-CDN improves user quality of experience and reduces cost of delivering content for the CPs.
Edge computing offers an additional layer of compute infrastructure closer to the data source before raw data from privacy-sensitive and performance-critical applications is transferred to a cloud data center. Deep Neural Networks (DNNs) are one class of applications that are reported to benefit from collaboratively computing between the edge and the cloud. A DNN is partitioned such that specific layers of the DNN are deployed onto the edge and the cloud to meet performance and privacy objectives. However, there is limited understanding of: (a) whether and how evolving operational conditions (increased CPU and memory utilization at the edge or reduced data transfer rates between the edge and the cloud) affect the performance of already deployed DNNs, and (b) whether a new partition configuration is required to maximize performance. A DNN that adapts to changing operational conditions is referred to as an adaptive DNN. This paper investigates whether there is a case for adaptive DNNs in edge computing by considering three questions: (i) Are DNNs sensitive to operational conditions? (ii) How sensitive are DNNs to operational conditions? (iii) Do individual or a combination of operational conditions equally affect DNNs? (iv) Is DNN partitioning sensitive to hardware architectures on the cloud/edge? The exploration is carried out in the context of 8 pre-trained DNN models and the results presented are from analyzing nearly 8 million data points. The results highlight that network conditions affects DNN performance more than CPU or memory related operational conditions. Repartitioning is noted to provide a performance gain in a number of cases, but a specific trend was not noted in relation to its correlation to the underlying hardware architecture. Nonetheless, the need for adaptive DNNs is confirmed.
Device-to-Device (D2D) communication can support the operation of cellular systems by reducing the traffic in the network infrastructure. In this paper, the benefits of D2D communication are investigated in the context of a Fog-Radio Access Network (F-RAN) that leverages edge caching and fronthaul connectivity for the purpose of content delivery. Assuming offline caching, out-of-band D2D communication, and an F-RAN with two edge nodes and two user equipments, an information-theoretically optimal caching and delivery strategy is presented that minimizes the delivery time in the high signal-to-noise ratio regime. The delivery time accounts for the latency caused by fronthaul, downlink, and D2D transmissions. The proposed optimal strategy is based on a novel scheme for an X-channel with receiver cooperation that leverages tools from real interference alignment. Insights are provided on the regimes in which D2D communication is beneficial.
In this work, we propose a content caching and delivery strategy to maximize throughput capacity in cache-enabled wireless networks. To this end, efficient betweenness (EB), which indicates the ratio of content delivery paths passing through a node, is first defined to capture the impact of content caching and delivery on network traffic load distribution. Aided by EB, throughput capacity is shown to be upper bounded by the minimal ratio of successful delivery probability (SDP) to EB among all nodes. Through effectively matching nodes EB with their SDP, the proposed strategy improves throughput capacity with low computation complexity. Simulation results show that the gap between the proposed strategy and the optimal one (obtained through exhausted search) is kept smaller than 6%.