scispace - formally typeset
Search or ask a question

Showing papers on "Cache published in 2021"


Journal ArticleDOI
TL;DR: The major purpose of this work is to create a novel and secure cache decision system (CDS) in a wireless network that operates over an SB, which will offer the users safer and efficient environment for browsing the Internet, sharing and managing large-scale data in the fog.
Abstract: This work proposes an innovative infrastructure of secure scenario which operates in a wireless-mobile 6G network for managing big data (BD) on smart buildings (SBs). Count on the rapid growth of telecommunication field new challenges arise. Furthermore, a new type of wireless network infrastructure, the sixth generation (6G), provides all the benefits of its past versions and also improves some issues which its predecessors had. In addition, relative technologies to the telecommunications filed, such as Internet of Things, cloud computing (CC) and edge computing (EC), can operate through a 6G wireless network. Take into account all these, we propose a scenario that try to combine the functions of the Internet of Things with CC, EC and BD in order to achieve a Smart and Secure environment. The major purpose of this work is to create a novel and secure cache decision system (CDS) in a wireless network that operates over an SB, which will offer the users safer and efficient environment for browsing the Internet, sharing and managing large-scale data in the fog. This CDS consisted of two types of servers, one cloud server and one edge server. In order to come up with our proposal, we study related cache scenarios systems which are listed, presented, and compared in this work.

229 citations


Journal ArticleDOI
TL;DR: Experimental results demonstrate that MPCF outperforms other baseline caching schemes in terms of the cache hit ratio in vehicular edge networks, and can greatly improve cache performance, effectively protect users’ privacy and significantly reduce communication costs.
Abstract: Content Caching at the edge of vehicular networks has been considered as a promising technology to satisfy the increasing demands of computation-intensive and latency-sensitive vehicular applications for intelligent transportation. The existing content caching schemes, when used in vehicular networks, face two distinct challenges: 1) Vehicles connected to an edge server keep moving, making the content popularity varying and hard to predict. 2) Cached content is easily out-of-date since each connected vehicle stays in the area of an edge server for a short duration. To address these challenges, we propose a Mobility-aware Proactive edge Caching scheme based on Federated learning (MPCF). This new scheme enables multiple vehicles to collaboratively learn a global model for predicting content popularity with the private training data distributed on local vehicles. MPCF also employs a Context-aware Adversarial AutoEncoder to predict the highly dynamic content popularity. Besides, MPCF integrates a mobility-aware cache replacement policy, which allows the network edges to add/evict contents in response to the mobility patterns and preferences of vehicles. MPCF can greatly improve cache performance, effectively protect users’ privacy and significantly reduce communication costs. Experimental results demonstrate that MPCF outperforms other baseline caching schemes in terms of the cache hit ratio in vehicular edge networks.

145 citations


Journal ArticleDOI
TL;DR: This study proposes an offloading model for a multi-user MEC system with multi-task, and an equivalent form of reinforcement learning is created where the state spaces are defined based on all possible solutions and the actions are defined on the basis of movement between the different states.
Abstract: Computation offloading at mobile edge computing (MEC) servers can mitigate the resource limitation and reduce the communication latency for mobile devices Thereby, in this study, we proposed an offloading model for a multi-user MEC system with multi-task In addition, a new caching concept is introduced for the computation tasks, where the application program and related code for the completed tasks are cached at the edge server Furthermore, an efficient model of task offloading and caching integration is formulated as a nonlinear problem whose goal is to reduce the total overhead of time and energy However, solving these types of problems is computationally prohibitive, especially for large-scale of mobile users Thus, an equivalent form of reinforcement learning is created where the state spaces are defined based on all possible solutions and the actions are defined on the basis of movement between the different states Afterwards, two effective Q-learning and Deep-Q-Network-based algorithms are proposed to derive the near-optimal solution for this problem Finally, experimental evaluations verify that our proposed model can substantially minimize the mobile devices’ overhead by deploying computation offloading and task caching strategy reasonably

115 citations


Journal ArticleDOI
TL;DR: This paper explores the cache deployment in a large-scale WiFi system, which contains 8,000 APs and serves more than 40,000 active users, to maximize the long-term caching gain, and proposes a cache deployment strategy, named LeaD, which is able to achieve the near-optimal caching performance and can outperform other benchmark strategies significantly.
Abstract: Widespread and large-scale WiFi systems have been deployed in many corporate locations, while the backhual capacity becomes the bottleneck in providing high-rate data services to a tremendous number of WiFi users. Mobile edge caching is a promising solution to relieve backhaul pressure and deliver quality services by proactively pushing contents to access points (APs). However, how to deploy cache in large-scale WiFi system is not well studied yet quite challenging since numerous APs can have heterogeneous traffic characteristics, and future traffic conditions are unknown ahead. In this paper, given the cache storage budget, we explore the cache deployment in a large-scale WiFi system, which contains 8,000 APs and serves more than 40,000 active users, to maximize the long-term caching gain. Specifically, we first collect two-month user association records and conduct intensive spatio-temporal analytics on WiFi traffic consumption, gaining two major observations. First, per AP traffic consumption varies in a rather wide range and the proportion of AP distributes evenly within the range, indicating that the cache size should be heterogeneously allocated in accordance to the underlying traffic demands. Second, compared to a single AP, the traffic consumption of a group of APs (clustered by physical locations) is more stable, which means that the short-term traffic statistics can be used to infer the future long-term traffic conditions. We then propose our cache deployment strategy, named LeaD (i.e., L arge-scale WiFi E dge c A che D eployment), in which we first cluster large-scale APs into well-sized edge nodes, then conduct the stationary testing on edge level traffic consumption and sample sufficient traffic statistics in order to precisely characterize long-term traffic conditions, and finally devise the TEG ( T raffic-w E ighted G reedy) algorithm to solve the long-term caching gain maximization problem. Extensive trace-driven experiments are carried out, and the results demonstrate that LeaD is able to achieve the near-optimal caching performance and can outperform other benchmark strategies significantly.

93 citations


Journal ArticleDOI
TL;DR: In this paper, the authors proposed a deep learning model to predict the contents need to be cached in self-driving cars and close proximity of selfdriving cars in multi-access edge computing servers attached to roadside units.
Abstract: Without steering wheel and driver’s seat, the self-driving cars will have new interior outlook and spaces that can be used for enhanced infotainment services. For traveling people, self-driving cars will be new places for engaging in infotainment services. Therefore, self-driving cars should determine themselves the infotainment contents that are likely to entertain their passengers. However, the choice of infotainment contents depends on passengers’ features such as age, emotion, and gender. Also, retrieving infotainment contents at data center can hinder infotainment services due to high end-to-end delay. To address these challenges, we propose infotainment caching in self-driving cars, where caching decisions are based on passengers’ features obtained using deep learning. First, we proposed deep learning models to predict the contents need to be cached in self-driving cars and close proximity of self-driving cars in multi-access edge computing servers attached to roadside units. Second, we proposed a communication model for retrieving infotainment contents to cache. Third, we proposed a caching model for retrieved contents. Fourth, we proposed a computation model for the cached contents, where cached contents can be served in different formats/qualities based on demands. Finally, we proposed an optimization problem whose goal is to link the proposed models into one optimization problem that minimizes the content downloading delay. To solve the formulated problem, a block successive majorization-minimization technique is applied. The simulation results show that the accuracy of prediction for the contents that need to be cached is 97.82% and our approach can minimize the delay.

86 citations


Journal ArticleDOI
TL;DR: This article proposes a lightweight sampling-based probabilistic approach, namely EDI-V, to help app vendors audit the integrity of their data cached on a large scale of edge servers, and proposes a new data structure named variable Merkle hash tree (VMHT) for generating the integrity proofs of those data replicas during the audit.
Abstract: Edge computing allows app vendors to deploy their applications and relevant data on distributed edge servers to serve nearby users. Caching data on edge servers can minimize users’ data retrieval latency. However, such cache data are subject to both intentional and accidental corruption in the highly distributed, dynamic, and volatile edge computing environment. Given a large number of edge servers and their limited computing resources, how to effectively and efficiently audit the integrity of app vendors’ cache data is a critical and challenging problem. This article makes the first attempt to tackle this Edge Data Integrity (EDI) problem. We first analyze the threat model and the audit objectives, then propose a lightweight sampling-based probabilistic approach, namely EDI-V, to help app vendors audit the integrity of their data cached on a large scale of edge servers. We propose a new data structure named variable Merkle hash tree (VMHT) for generating the integrity proofs of those data replicas during the audit. VMHT can ensure the audit accuracy of EDI-V by maintaining sampling uniformity. EDI-V allows app vendors to inspect their cache data and locate the corrupted ones efficiently and effectively. Both theoretical analysis and comprehensively experimental evaluation demonstrate the efficiency and effectiveness of EDI-V.

85 citations


Journal ArticleDOI
TL;DR: The simulation results prove that the proposed density-based content distribution method can obviously reduce the average transmission delay of content distribution under different network conditions and has better stability and self-adaptability under continuous time variation.
Abstract: The satellite-terrestrial networks (STN) utilize the spacious coverage and low transmission latency of Low Earth Orbit (LEO) constellation to distribute requested content for ground subscribers. With the development of storage and computing capacity of satellite onboard equipment, it is considered promising to leverage in-network caching technology on STN to improve content distribution efficiency. However, traditional ground network caching schemes are not suitable in STN, considering dynamic satellite propagation and time-varying topology. More specifically, the unevenness of user distribution results in difficulties for assurance of quality of experience. To address these problems, we firstly propose a density-based block division algorithm to divide the content subscribers into a series of blocks with different sizes according to user density. The LEO satellite orbit and time-varying network model is established to describe STN topology. Next, we propose an approximate minimum coverage vertex set algorithm and a novel cache node selection algorithm for optimal user blocks matching. The simulation results prove that the proposed density-based content distribution method can obviously reduce the average transmission delay of content distribution under different network conditions and has better stability and self-adaptability under continuous time variation.

81 citations


Journal ArticleDOI
TL;DR: An attention-weighted federated deep reinforcement learning (AWFDRL) model that uses federated learning to improve the training efficiency of the Q-learning network by considering the limited computing and storage capacity, and incorporates an attention mechanism to optimize the aggregation weights to avoid the imbalance of local model quality is designed.
Abstract: In order to meet the growing demands for multimedia service access and release the pressure of the core network, edge caching and device-to-device (D2D) communication have been regarded as two promising techniques in next generation mobile networks and beyond. However, most existing related studies lack consideration of effective cooperation and adaptability to the dynamic network environments. In this article, based on the flexible trilateral cooperation among user equipment, edge base stations and a cloud server, we propose a D2D-assisted heterogeneous collaborative edge caching framework by jointly optimizing the node selection and cache replacement in mobile networks. We formulate the joint optimization problem as a Markov decision process, and use a deep Q-learning network to solve the long-term mixed integer linear programming problem. We further design an attention-weighted federated deep reinforcement learning (AWFDRL) model that uses federated learning to improve the training efficiency of the Q-learning network by considering the limited computing and storage capacity, and incorporates an attention mechanism to optimize the aggregation weights to avoid the imbalance of local model quality. We prove the convergence of the corresponding algorithm, and present simulation results to show the effectiveness of the proposed AWFDRL framework in reducing average delay of content access, improving hit rate and offloading traffic.

71 citations


Journal ArticleDOI
TL;DR: A comprehensive taxonomy of machine learning techniques for in-network caching in edge networks is formulated and a comparative analysis of the state-of-the-art literature is presented with respect to the parameters identified in the taxonomy.

71 citations


Journal ArticleDOI
TL;DR: Wang et al. as discussed by the authors proposed a new algorithm in which blockchain assisted Compressed algoRithm of fEderated leArning is applied for conTent caching, called CREAT to predict cached files.
Abstract: Edge computing architectures can help us quickly process the data collected by Internet of Things (IoT) and caching files to edge nodes can speed up the response speed of IoT devices requesting files. Blockchain architectures can help us ensure the security of data transmitted by IoT. Therefore, we have proposed a system which combines IoT devices, edge nodes, remote cloud and blockchain. In the system, we designed a new algorithm in which blockchain-assisted Compressed algoRithm of fEderated leArning is applied for conTent caching, called CREAT to predict cached files. In CREAT algorithm, each edge node uses local data to train a model and then uses the model to learn the features of users and files, so as to predict popular files to improve cache hit rate. In order to ensure the security of edge nodes’ data, we use federated learning (FL) to enable multiple edge nodes to cooperate in training without sharing data. In addition, for the purpose of reducing communication load in FL, we will compress gradients uploaded by edge nodes to reduce the time required for communication. What’s more, in order to ensure the security of the data transmitted in CREAT algorithm, we have incorporated blockchain technology in the algorithm. We design four smart contracts for decentralized entities to record and verify the transactions to ensure the security of data. We used MovieLens data sets for experiments and we can see that CREAT greatly improves the cache hit rate and reduces the time required to upload data.

64 citations


Proceedings ArticleDOI
06 Jun 2021
TL;DR: In this article, the long-range history context is distilled into an augmented memory bank to reduce self-attention's computation complexity, and a cache mechanism saves the computation for the key and value in selfattention for the left context.
Abstract: This paper proposes an efficient memory transformer Emformer for low latency streaming speech recognition. In Emformer, the long-range history context is distilled into an augmented memory bank to reduce self-attention’s computation complexity. A cache mechanism saves the computation for the key and value in self-attention for the left context. Emformer applies a parallelized block processing in training to support low latency models. We carry out experiments on benchmark LibriSpeech data. Under average latency of 960 ms, Emformer gets WER 2.50% on test-clean and 5.62% on test-other. Comparing with a strong baseline augmented memory transformer (AM-TRF), Emformer gets 4.6 folds training speedup and 18% relative real-time factor (RTF) reduction in decoding with relative WER reduction 17% on test-clean and 9% on test-other. For a low latency scenario with an average latency of 80 ms, Emformer achieves WER 3.01% on test-clean and 7.09% on test-other. Comparing with the LSTM baseline with the same latency and model size, Emformer gets relative WER reduction 9% and 16% on test-clean and test-other, respectively.

Journal ArticleDOI
TL;DR: The challenges of NDN-IoT caching are identified with the aim to develop a new hybrid strategy for efficient data delivery and it is observed that the proposed hybrid strategy outperformed to achieve a higher caching performance ofNDN-based IoT scenarios.
Abstract: Internet of Things (IoT) and Named Data Network (NDN) are innovative technologies to meet up the future Internet requirements. NDN is considered as an enabling approach to improving data dissemination in IoT scenarios. NDN delivers in-network caching, which is the most prominent feature to provide fast data dissemination as compared to Internet Protocol (IP) based communication. The proper integration of caching placement strategies and replacement policies is the most suitable approach to support IoT networks. It can improve multicast communication which minimizes the delay in responding to IoT-based environments. Besides, these approaches are playing a most significant role in increasing the overall performance of NDN-based IoT networks. To this end, in this paper, the challenges of NDN-IoT caching are identified with the aim to develop a new hybrid strategy for efficient data delivery. The proposed strategy is comparatively and extensively studied with NDN-IoT caching strategies through an extensive simulation in terms of average latency, cache hit ratio and average stretch ratio. From the simulation findings, it is observed that the proposed hybrid strategy outperformed to achieve a higher caching performance of NDN-based IoT scenarios.

Journal ArticleDOI
TL;DR: In this paper, the authors present a real-time neural radiance caching method for path-traced global illumination, which makes no assumptions about the lighting, geometry, and materials.
Abstract: We present a real-time neural radiance caching method for path-traced global illumination. Our system is designed to handle fully dynamic scenes, and makes no assumptions about the lighting, geometry, and materials. The data-driven nature of our approach sidesteps many difficulties of caching algorithms, such as locating, interpolating, and updating cache points. Since pretraining neural networks to handle novel, dynamic scenes is a formidable generalization challenge, we do away with pretraining and instead achieve generalization via adaptation, i.e. we opt for training the radiance cache while rendering. We employ self-training to provide low-noise training targets and simulate infinite-bounce transport by merely iterating few-bounce training updates. The updates and cache queries incur a mild overhead---about 2.6ms on full HD resolution---thanks to a streaming implementation of the neural network that fully exploits modern hardware. We demonstrate significant noise reduction at the cost of little induced bias, and report state-of-the-art, real-time performance on a number of challenging scenarios.

Proceedings ArticleDOI
19 Apr 2021
TL;DR: In this paper, cache coherence is used instead of virtual memory for tracking applications' memory accesses transparently, at cache-line granularity, eliminating page faults from the application critical path when accessing remote data, and decoupling the application memory access tracking from the virtual memory page size.
Abstract: Disaggregated memory can address resource provisioning inefficiencies in current datacenters. Multiple software runtimes for disaggregated memory have been proposed in an attempt to make disaggregated memory practical. These systems rely on the virtual memory subsystem to transparently offer disaggregated memory to applications using a local memory abstraction. Unfortunately, using virtual memory for disaggregation has multiple limitations, including high overhead that comes from the use of page faults to identify what data to fetch and cache locally, and high dirty data amplification that comes from the use of page-granularity for tracking changes to the cached data (4KB or higher). In this paper, we propose a fundamentally new approach to designing software runtimes for disaggregated memory that addresses these limitations. Our main observation is that we can use cache coherence instead of virtual memory for tracking applications' memory accesses transparently, at cache-line granularity. This simple idea (1) eliminates page faults from the application critical path when accessing remote data, and (2) decouples the application memory access tracking from the virtual memory page size, enabling cache-line granularity dirty data tracking and eviction. Using this observation, we implemented a new software runtime for disaggregated memory that improves average memory access time by 1.7-5X and reduces dirty data amplification by 2-10X, compared to state-of-the-art systems.

Journal ArticleDOI
TL;DR: Wang et al. as discussed by the authors studied how to efficiently offload dependent tasks to edge nodes with limited (and predetermined) service caching, and designed an efficient convex programming based algorithm (CP) to solve this problem.
Abstract: In Mobile Edge Computing (MEC), many tasks require specific service support for execution and in addition, have a dependent order of execution among the tasks. However, previous works often ignore the impact of having limited services cached at the edge nodes on (dependent) task offloading, thus may lead to an infeasible offloading decision or a longer completion time. To bridge the gap, this article studies how to efficiently offload dependent tasks to edge nodes with limited (and predetermined) service caching. We formally define the problem of offloading dependent tasks with service caching (ODT-SC), and prove that there exists no algorithm with constant approximation for this hard problem. Then, we design an efficient convex programming based algorithm (CP) to solve this problem. Moreover, we study a special case with a homogeneous MEC and propose a favorite successor based algorithm (FS) to solve this special case with a competitive ratio of $O(1)$ O ( 1 ) . Extensive simulation results using Google data traces show that our proposed algorithms can significantly reduce applications’ completion time by about 21-47 percent compared with other alternatives.

Proceedings ArticleDOI
09 Aug 2021
TL;DR: In this article, the authors present measurement and insights for Linux kernel network stack performance for 100Gbps access link bandwidths and find that such high bandwidth links, coupled with relatively stagnant technology trends for other host resources (e.g., CPU speeds and capacity, cache sizes, NIC buffer sizes, etc.), mark a fundamental shift in host network stack bottlenecks.
Abstract: Traditional end-host network stacks are struggling to keep up with rapidly increasing datacenter access link bandwidths due to their unsustainable CPU overheads. Motivated by this, our community is exploring a multitude of solutions for future network stacks: from Linux kernel optimizations to partial hardware offload to clean-slate userspace stacks to specialized host network hardware. The design space explored by these solutions would benefit from a detailed understanding of CPU inefficiencies in existing network stacks. This paper presents measurement and insights for Linux kernel network stack performance for 100Gbps access link bandwidths. Our study reveals that such high bandwidth links, coupled with relatively stagnant technology trends for other host resources (e.g., CPU speeds and capacity, cache sizes, NIC buffer sizes, etc.), mark a fundamental shift in host network stack bottlenecks. For instance, we find that a single core is no longer able to process packets at line rate, with data copy from kernel to application buffers at the receiver becoming the core performance bottleneck. In addition, increase in bandwidth-delay products have outpaced the increase in cache sizes, resulting in inefficient DMA pipeline between the NIC and the CPU. Finally, we find that traditional loosely-coupled design of network stack and CPU schedulers in existing operating systems becomes a limiting factor in scaling network stack performance across cores. Based on insights from our study, we discuss implications to design of future operating systems, network protocols, and host hardware.

Journal ArticleDOI
TL;DR: This article proposes an MEC service pricing scheme to coordinate with the service caching decisions and control WDs’ task offloading behavior in a cellular network and derives the optimal threshold-based offloading policy that can be easily adopted by the WDs in Stage II at the Bayesian equilibrium.
Abstract: Provided with mobile edge computing (MEC) services, wireless devices (WDs) no longer have to experience long latency in running their desired programs locally, but can pay to offload computation tasks to the edge server. Given its limited storage space, it is important for the edge server at the base station (BS) to determine which service programs to cache by meeting and guiding WDs’ offloading decisions. In this article, we propose an MEC service pricing scheme to coordinate with the service caching decisions and control WDs’ task offloading behavior in a cellular network. We propose a two-stage dynamic game of incomplete information to model and analyze the two-stage interaction between the BS and multiple associated WDs. Specifically, in Stage I, the BS determines the MEC service caching and announces the service program prices to the WDs, with the objective to maximize its expected profit under both storage and computation resource constraints. In Stage II, given the prices of different service programs, each WD selfishly decides its offloading decision to minimize individual service delay and cost, without knowing the other WDs’ desired program types or local execution delays. Despite the lack of WD’s information and the coupling of all the WDs’ offloading decisions, we derive the optimal threshold-based offloading policy that can be easily adopted by the WDs in Stage II at the Bayesian equilibrium. In particular, a WD is more likely to offload when there are fewer WDs competing for the edge server’s computation resource, or when it perceives a good channel condition or low MEC service price. Then, by predicting the WDs’ offloading equilibrium, we jointly optimize the BS’ pricing and service caching in Stage I via a low-complexity algorithm. In particular, we first study the differentiated pricing scheme and prove that the same price should be charged to the cached programs of the same workload. Motivated by this analysis, we further propose a low-complexity uniform pricing heuristics.

Journal ArticleDOI
TL;DR: In this paper, the authors present a real-time neural radiance caching method for path-traced global illumination, which makes no assumptions about the lighting, geometry, and materials.
Abstract: We present a real-time neural radiance caching method for path-traced global illumination. Our system is designed to handle fully dynamic scenes, and makes no assumptions about the lighting, geometry, and materials. The data-driven nature of our approach sidesteps many difficulties of caching algorithms, such as locating, interpolating, and updating cache points. Since pretraining neural networks to handle novel, dynamic scenes is a formidable generalization challenge, we do away with pretraining and instead achieve generalization via adaptation, i.e. we opt for training the radiance cache while rendering. We employ self-training to provide low-noise training targets and simulate infinite-bounce transport by merely iterating few-bounce training updates. The updates and cache queries incur a mild overhead -- about 2.6ms on full HD resolution -- thanks to a streaming implementation of the neural network that fully exploits modern hardware. We demonstrate significant noise reduction at the cost of little induced bias, and report state-of-the-art, real-time performance on a number of challenging scenarios.

Proceedings ArticleDOI
23 May 2021
TL;DR: This paper consolidates existing randomization-based secure caches into a generic cache model, and comprehensively analyze the security of existing designs, including CEASER-S and SCATTERCACHE, by mapping them to instances of this model.
Abstract: Recent secure cache designs aim to mitigate side-channel attacks by randomizing the mapping from memory addresses to cache sets. As vendors investigate deployment of these caches, it is crucial to understand their actual security.In this paper, we consolidate existing randomization-based secure caches into a generic cache model. We then comprehensively analyze the security of existing designs, including CEASER-S and SCATTERCACHE, by mapping them to instances of this model. We tailor cache attacks for randomized caches using a novel PRIME+PRUNE+PROBE technique, and optimize it using burst accesses, bootstrapping, and multi-step profiling. PRIME+ PRUNE+PROBE constructs probabilistic but reliable eviction sets, enabling attacks previously assumed to be computationally infeasible. We also simulate an end-to-end attack, leaking secrets from a vulnerable AES implementation. Finally, a case study of CEASER-S reveals that cryptographic weaknesses in the randomization algorithm can lead to a complete security subversion.Our systematic analysis yields more realistic and comparable security levels for randomized caches. As we quantify how design parameters influence the security level, our work leads to important conclusions for future work on secure cache designs.

Journal ArticleDOI
TL;DR: This article proposes a social-aware vehicular edge caching mechanism that dynamically orchestrates the cache capability of roadside units (RSUs) and smart vehicles according to user preference similarity and service availability and proposes deep learning empowered optimal caching schemes.
Abstract: The rapid proliferation of smart vehicles along with the advent of powerful applications bring stringent requirements on massive content delivery. Although vehicular edge caching can facilitate delay-bounded content transmission, constrained storage capacity and limited serving range of an individual cache server as well as highly dynamic topology of vehicular networks may degrade the efficiency of content delivery. To address the problem, in this article, we propose a social-aware vehicular edge caching mechanism that dynamically orchestrates the cache capability of roadside units (RSUs) and smart vehicles according to user preference similarity and service availability. Furthermore, catering to the complexity and variability of vehicular social characteristics, we leverage the digital twin technology to map the edge caching system into virtual space, which facilitates constructing the social relation model. Based on the social model, a new concept of vehicular cache cloud is developed to incorporate the correlation of content storing between multiple cache-enabled vehicles in diverse traffic environments. Then, we propose deep learning empowered optimal caching schemes, jointly considering the social model construction, cache cloud formation, and cache resource allocation. We evaluate the proposed schemes based on real traffic data. Numerical results demonstrate that our edge caching schemes have great advantages in optimizing caching utility.

Journal ArticleDOI
TL;DR: This article investigates the issue of secure transmission in a cache-enabled UAV-relaying network with D2D communications in the presence of an eavesdropper, and proposes an alternating iterative algorithm based on the block alternating descent and successive convex approximation methods to solve the problem.
Abstract: With the exponential growth of data traffic, the use of caching and device-to-device (D2D) communication has been recognized as an effective approach for mitigating the backhaul bottleneck in unmanned aerial vehicle (UAV)-assisted networks. In this article, we investigate the issue of secure transmission in a cache-enabled UAV-relaying network with D2D communications in the presence of an eavesdropper. Specifically, both UAVs and D2D users are equipped with cache memory, which can prestore some popular content to collaboratively serve users. Considering the fairness among users, we formulate an optimization problem to maximize the minimum secrecy rate among users, by jointly optimizing the user association and UAV scheduling, transmission power, and UAV trajectory over a finite period. The joint design problem is a nonconvex mixed-integer programming problem. To efficiently solve this problem, we propose an alternating iterative algorithm based on the block alternating descent and successive convex approximation methods. Specifically, the user association and UAV scheduling, UAV trajectory, and transmission power are optimized alternately in each iteration, and the convergence of the algorithm is proven. Extensive numerical results show that the proposed joint design scheme significantly outperforms other benchmark schemes in terms of the secrecy rate.

Journal ArticleDOI
TL;DR: This paper proposes a Cooperative Caching scheme based on Social Attributes and Mobility Prediction (CCSAMP) for VCCN, based on the observation that vehicles move around and are liable to contact each other according to drivers’ common interests or social similarities, which has higher cache hit ratio and lower content access delay compared to other state-of-the-art schemes.
Abstract: Communications in vehicular ad-hoc network (VANET) are subject to performance degradation as results of channel fading and intermittent network connectivity. The emerging Vehicular Content Centric Network (VCCN) is promising in supporting the needs of contents and alleviating the communication problems in VANET. Specifically, to improve the cache hit ratio and reduce the access delay of content retrieval, it helps to choose the appropriate vehicles to cache the frequently accessed data items. In this paper, we propose a Cooperative Caching scheme based on Social Attributes and Mobility Prediction (CCSAMP) for VCCN. CCSAMP is based on the observation that vehicles move around and are liable to contact each other according to drivers’ common interests or social similarities. A caching node sharing more social attributes with the content requester is more likely to be interested in the same contents and distribute the contents to others with similar interests. Furthermore, a caching node that frequently meets other nodes is a better candidate to keep cache copies. To increase the network performance, CCSAMP also exploits the regularity of vehicle moving behaviors to predict the chance for a vehicle to reach hot zones based on Hidden Markov Model (HMM). We evaluate CCSAMP through the ONE simulator to demonstrate its higher cache hit ratio and lower content access delay compared to other state-of-the-art schemes.

Proceedings ArticleDOI
21 Apr 2021
TL;DR: In this paper, a transparent, vertically and horizontally elastic in-memory caching system for FaaS platforms, distributed over the worker nodes, is introduced, using machine learning models adjusted for typical function input data categories (e.g., multimedia formats).
Abstract: Cloud applications based on the "Functions as a Service" (FaaS) paradigm have become very popular. Yet, due to their stateless nature, they must frequently interact with an external data store, which limits their performance. To mitigate this issue, we introduce OFC, a transparent, vertically and horizontally elastic in-memory caching system for FaaS platforms, distributed over the worker nodes. OFC provides these benefits cost-effectively by exploiting two common sources of resource waste: (i) most cloud tenants overprovision the memory resources reserved for their functions because their footprint is non-trivially input-dependent and (ii) FaaS providers keep function sandboxes alive for several minutes to avoid cold starts. Using machine learning models adjusted for typical function input data categories (e.g., multimedia formats), OFC estimates the actual memory resources required by each function invocation and hoards the remaining capacity to feed the cache. We build our OFC prototype based on enhancements to the OpenWhisk FaaS platform, the Swift persistent object store, and the RAM-Cloud in-memory store. Using a diverse set of workloads, we show that OFC improves by up to 82 % and 60 % respectively the execution time of single-stage and pipelined functions.

Journal ArticleDOI
TL;DR: A cache updating system with a source, a cache and a user, an alternating maximization based method to find the update rates for the cache and for the user is provided to maximize the freshness of the files at the user.
Abstract: We consider a cache updating system with a source, a cache and a user. There are $n$ files. The source keeps the freshest version of the files which are updated with known rates $\lambda _{i}$ . The cache downloads and keeps the freshest version of the files from the source with rates $c_{i}$ . The user gets updates from the cache with rates $u_{i}$ . When the user gets an update, it either gets a fresh update from the cache or the file at the cache becomes outdated by a file update at the source in which case the user gets an outdated update. We find an analytical expression for the average freshness of the files at the user. Next, we generalize our setting to the case where there are multiple caches in between the source and the user, and find the average freshness at the user. We provide an alternating maximization based method to find the update rates for the cache(s), $c_{i}$ , and for the user, $u_{i}$ , to maximize the freshness of the files at the user. We observe that for a given set of update rates for the user (resp. for the cache), the optimal rate allocation policy for the cache (resp. for the user) is a threshold policy , where the optimal update rates for rapidly changing files at the source may be equal to zero. Finally, we consider a system where multiple users are connected to a single cache and find update rates for the cache and the users to maximize the total freshness over all users.

Journal ArticleDOI
TL;DR: A hybrid human-artificial intelligence approach is developed to improve the user hit rate for video caching and guarantees the user fairness in terms of video coding rate under statistical delay constraint and edge caching capacity constraint.
Abstract: In this paper, a video service enhancement strategy is investigated under an edge-cloud collaboration framework, where video caching and delivery decisions are made at the cloud and edge respectively. We aim to guarantee the user fairness in terms of video coding rate under statistical delay constraint and edge caching capacity constraint. A hybrid human-artificial intelligence approach is developed to improve the user hit rate for video caching. Specifically, individual user interest is first characterized by merging factorization machine (FM) model and multi-layer perceptron (MLP) model, where both low-order and high-order features can be well learned simultaneously. Thereafter, a social aware similarity model is constructed to transfer individual user interest to group interest, based on which, videos can be selected to cache at the network edge. Furthermore, a dual bisection exploration scheme is proposed to optimize wireless resource allocation and video coding rate. The effectiveness of the proposed video caching and delivery scheme is finally validated by extensive experiments with a real-world dataset.

Journal ArticleDOI
TL;DR: A strategy that uses reinforcement learning algorithm to optimize cache schemes on different devices to maximize the efficiency of content cache is designed and can enhance the cache hit ratio by 10%-20% compared with the well-known counterparts.
Abstract: The rapid development of 6G can help to bring autonomous driving closed to the reality. Drivers and passengers will have more time for work and leisure spending in the vehicles, further generating a lot of data requirements. However, edge resources from small base stations are insufficient to match the wide variety of services of the future vehicular networks. Besides, due to the high-speed nature of the vehicles, users have to switch the connections among different base stations, whereas such way will cause external latency during the data request. Therefore, it is vital to enable the local cache of vehicle users to realize the reliable autonomous driving. In this paper, we consider caching the contents in the local cache, small base station, and edge server. In practice, the request preference of some single users may be different from a whole region. To maximize the efficiency of content cache, we design a strategy that uses reinforcement learning algorithm to optimize cache schemes on different devices. The experimental results demonstrate that our strategy can enhance the cache hit ratio by 10%-20% compared with the well-known counterparts.

Journal ArticleDOI
TL;DR: Deep reinforcement learning-based routing (DRL-R) is proposed, a method that recombines multiple network resources with different metrics, where it recombine cache and bandwidth by quantifying their contribution score in reducing the delay.

Journal ArticleDOI
TL;DR: In this article, a clustering-based long short-term memory (C-LTSM) approach was proposed to predict the number of content requests using historical request information.
Abstract: Coded caching is effective in leveraging the accumulated storage size in wireless networks by distributing different coded segments of each file in multiple cache nodes. This paper aims to find a wireless coded caching policy to minimize the total discounted network cost, which involves both transmission delay and cache replacement cost, using tools from deep learning. The problem is known to be challenging due to the unknown, time-variant content popularity as well as the continuous, high-dimensional action space. We first propose a clustering based long short-term memory (C-LTSM) approach to predict the number of content requests using historical request information. This approach exploits the correlation of the historical request information between different files through clustering. Based on the predicted results, we then propose a supervised deep deterministic policy gradient (SDDPG) approach. This approach, on one hand, can learn the caching policy in continuous action space by using the actor-critic architecture. On the other hand, it accelerates the learning process by pre-training the actor network based on the solution of an approximate problem that minimizes the per-slot cost. Real-world trace-based numerical results show that the proposed prediction and caching policy using deep learning outperform the considered existing methods.

Journal ArticleDOI
TL;DR: A low-latency virtual reality (VR) delivery system where an unmanned aerial vehicle (UAV) base station (U-BS) is deployed to deliver VR content from a cloud server to multiple ground VR users, which indicates that caching is helpful to reduce latency.
Abstract: In this paper, we propose a low-latency virtual reality (VR) delivery system where an unmanned aerial vehicle (UAV) base station (U-BS) is deployed to deliver VR content from a cloud server to multiple ground VR users. Each VR input data requested by the VR users can be either projected at the U-BS before transmission or processed locally at each user. Popular VR input data is cached at the U-BS to further reduce backhaul latency from the cloud server. For this system, we design a low-complexity iterative algorithm to minimize the maximum communications and computing latency among all VR users subject to the computing, caching and transmit power constraints, which is guaranteed to converge. Numerical results indicate that our proposed algorithm can achieve a lower latency compared to other benchmark schemes. Moreover, we observe that the maximum latency mainly comes from communication latency when the bandwidth resource is limited, while it is dominated by computing latency when computing capacity is low. In addition, we find that caching is helpful to reduce latency.

Journal ArticleDOI
TL;DR: A new caching algorithm, called Similarity-Aware Popularity-based Caching (SAPoC), is presented in this paper to promote the performance of wireless edge caching in dynamic scenarios through utilizing the similarity among contents.