scispace - formally typeset
Search or ask a question

Showing papers on "Service-level agreement published in 2020"


Journal ArticleDOI
TL;DR: A virtual machine consolidation algorithm with multiple usage prediction (VMCUP-M) to improve the energy efficiency of cloud data centers and reduces the number of migrations and the power consumption of the servers while complying with the service level agreement.
Abstract: Virtual machine consolidation aims at reducing the number of active physical servers in a data center so as to decrease the total power consumption. In this context, most of the existing solutions rely on aggressive virtual machine migration, thus resulting in unnecessary overhead and energy wastage. Besides, virtual machine consolidation should take into account multiple resource types at the same time, since CPU is not the only critical resource in cloud data centers. In fact, also memory and network bandwidth can become a bottleneck, possibly causing violations in the service level agreement. This article presents a virtual machine consolidation algorithm with multiple usage prediction (VMCUP-M) to improve the energy efficiency of cloud data centers. In this context, multiple usage refers to both resource types and the horizon employed to predict future utilization. Our algorithm is executed during the virtual machine consolidation process to estimate the long-term utilization of multiple resource types based on the local history of the considered servers. The joint use of current and predicted resource utilization allows for a reliable characterization of overloaded and underloaded servers, thereby reducing both the load and the power consumption after consolidation. We evaluate our solution through simulations on both synthetic and real-world workloads. The obtained results show that consolidation with multiple usage prediction reduces the number of migrations and the power consumption of the servers while complying with the service level agreement.

126 citations


Journal ArticleDOI
TL;DR: In this paper, the authors proposed Generative Adversarial Network-powered Deep Distributional Q Network (GAN-DDQN) to learn the action-value distribution driven by minimizing the discrepancy between the estimated action value distribution and the target action value distributions.
Abstract: Network slicing is a key technology in 5G communications system. Its purpose is to dynamically and efficiently allocate resources for diversified services with distinct requirements over a common underlying physical infrastructure. Therein, demand-aware resource allocation is of significant importance to network slicing. In this paper, we consider a scenario that contains several slices in a radio access network with base stations that share the same physical resources (e.g., bandwidth or slots). We leverage deep reinforcement learning (DRL) to solve this problem by considering the varying service demands as the environment state and the allocated resources as the environment action . In order to reduce the effects of the annoying randomness and noise embedded in the received service level agreement (SLA) satisfaction ratio (SSR) and spectrum efficiency (SE), we primarily propose generative adversarial network-powered deep distributional Q network (GAN-DDQN) to learn the action-value distribution driven by minimizing the discrepancy between the estimated action-value distribution and the target action-value distribution. We put forward a reward-clipping mechanism to stabilize GAN-DDQN training against the effects of widely-spanning utility values. Moreover, we further develop Dueling GAN-DDQN, which uses a specially designed dueling generator, to learn the action-value distribution by estimating the state-value distribution and the action advantage function. Finally, we verify the performance of the proposed GAN-DDQN and Dueling GAN-DDQN algorithms through extensive simulations.

109 citations


Journal ArticleDOI
TL;DR: A SLA-aware autonomic resource management technique called STAR which mainly focuses on reducing SLA violation rate for the efficient delivery of cloud services and optimizing other QoS parameters which effect efficient cloud service delivery is presented.
Abstract: Cloud computing has recently emerged as an important service to manage applications efficiently over the Internet. Various cloud providers offer pay per use cloud services that requires Quality of Service (QoS) management to efficiently monitor and measure the delivered services through Internet of Things (IoT) and thus needs to follow Service Level Agreements (SLAs). However, providing dedicated cloud services that ensure user's dynamic QoS requirements by avoiding SLA violations is a big challenge in cloud computing. As dynamism, heterogeneity and complexity of cloud environment is increasing rapidly, it makes cloud systems insecure and unmanageable. To overcome these problems, cloud systems require self-management of services. Therefore, there is a need to develop a resource management technique that automatically manages QoS requirements of cloud users thus helping the cloud providers in achieving the SLAs and avoiding SLA violations. In this paper, we present SLA-aware autonomic resource management technique called STAR which mainly focuses on reducing SLA violation rate for the efficient delivery of cloud services. The performance of the proposed technique has been evaluated through cloud environment. The experimental results demonstrate that STAR is efficient in reducing SLA violation rate and in optimizing other QoS parameters which effect efficient cloud service delivery.

87 citations


Journal ArticleDOI
TL;DR: A profit-aware application placement policy is proposed using constraint Integer Linear Programming model that simultaneously enhances profit and ensures QoS during application placement on computing instances and provides compensation to users for any violation of Service Level Agreement.

81 citations


Journal ArticleDOI
TL;DR: A blockchain-based secure data processing framework for an edge envisioned V2X environment (hereafter referred to as BloCkEd), which comprises an optimal container-based data processing scheme, and a blockchain- based data integrity management scheme, designed to minimize link breakage and reducing latency.
Abstract: There has been an increasing trend of moving computing activities closer to the edge of the network, particularly in smart city applications (e.g., vehicle-to-everything – V2X). Such a paradigm allows the end user’s requests to be handled/processed by nodes at the edge of the network; thus, reducing latency, and preserving privacy of user data/activities. However, there are a number of challenges in such an edge computing ecosystem. Examples include (1) potential inappropriate utilization of resources at the edge nodes, (2) operational challenges in cache management and data integrity due to data migration between edge nodes, particularly when dealing with vehicular mobility in a V2X application, and (3) high energy consumption due to continuous link breakage and subsequent reestablishment of link(s). Therefore in this paper, we design a blockchain-based secure data processing framework for an edge envisioned V2X environment (hereafter referred to as BloCkEd ). Specifically, a multi-layered edge-enabled V2X system model for BloCkEd is presented, which includes the formulation of a multi-objective optimization problem. In addition, BloCkEd comprises an optimal container-based data processing scheme, and a blockchain-based data integrity management scheme, designed to minimize link breakage and reducing latency. Using Chandigarh City, India, as the scenario, we implement and evaluate the proposed approach in terms of its latency, energy consumption, and service level agreement compliance.

69 citations


Journal ArticleDOI
TL;DR: A Power-Aware technique depending on Particle Swarm Optimization (PAPSO) to determine the near-optimal placement for the migrated VMs and the experimental results show that PAPSO does not violate SLA and outperforms the Power- aware Best Fit Decreasing algorithm (PABFD).
Abstract: With the widespread usage of cloud computing to benefit from its services, cloud service providers have invested in constructing large scale data centers. Consequently, a tremendous increase in energy consumption has arisen in conjunction with its results, including a remarkable rise in costs of operating and cooling servers. Besides, increasing energy consumption has a significant impact on the environment due to emissions of carbon dioxide. Dynamic consolidation of Virtual Machines (VMs) into the minimal number of Physical Machines (PMs) is considered as one of the magic solutions to manage power consumption. The virtual machine placement problem is a critical issue for good VM consolidation. This paper proposes a Power-Aware technique depending on Particle Swarm Optimization (PAPSO) to determine the near-optimal placement for the migrated VMs. A discrete version of Particle Swarm Optimization (PSO) is adopted based on a decimal encoding to map the migrated VMs to the best appropriate PMs. Furthermore, an effective minimization fitness function is employed to reduce power consumption without violating the Service Level Agreement (SLA). Specifically, PAPSO consolidates the migrated VMs into the minimum number of PMs with a major constraint to decrease the number of overloaded hosts as much as possible. Therefore, the number of VM migrations can be reduced drastically by taking into consideration the main sources for VM migrations; overloaded hosts and underloaded ones. PAPSO is implemented in CloudSim and the experimental results on random workloads with different sizes of VMs and PMs show that PAPSO does not violate SLA and outperforms the Power-Aware Best Fit Decreasing algorithm (PABFD). It can reduce about 8.01%, 39.65%, 66.33%, and 11.87% on average in terms of consumed energy, number of VM migrations, number of host shutdowns and the combined metric Energy SLA Violation (ESV), respectively.

67 citations


Journal ArticleDOI
01 Oct 2020
TL;DR: An improved and adaptive differential evolution algorithm is developed to improve the learning efficiency of predictive model and is capable of optimizing the best suitable mutation operator and crossover operator.
Abstract: Cloud computing promises elasticity, flexibility and cost-effectiveness to satisfy service level agreement conditions. The cloud service providers should plan and provision the computing resources rapidly to ensure the availability of infrastructure to match the demands with closed proximity. The workload prediction has become critical as it can be helpful in managing the infrastructure effectively. In this paper, we present a workload forecasting framework based on neural network model with supervised learning technique. An improved and adaptive differential evolution algorithm is developed to improve the learning efficiency of predictive model. The algorithm is capable of optimizing the best suitable mutation operator and crossover operator. The prediction accuracy and convergence rate of the learning are observed to be improved due to its adaptive behavior in pattern learning from sampled data. The predictive model’s performance is evaluated on four real-world data traces including Google cluster trace and NASA Kennedy Space Center logs. The results are compared with state-of-the-art methods, and improvements up to 91%, 97% and 97.2% are observed over self-adaptive differential evolution, backpropagation and average-based workload prediction techniques, respectively.

50 citations


Journal ArticleDOI
TL;DR: This paper addresses the issue of resource provisioning as an enabler for end-to-end dynamic slicing in software defined networking/network function virtualization (SDN/NFV)-based fifth generation (5G) networks, and highlights the role of the underlying hyperparameters in the trade-off between overprovisioning and slices’ isolation.
Abstract: In this paper, we address the issue of resource provisioning as an enabler for end-to-end dynamic slicing in software defined networking/network function virtualization (SDN/NFV)-based fifth generation (5G) networks. The different slices’ tenants (i.e. logical operators) are dynamically allocated isolated portions of physical resource blocks (PRBs), baseband processing resources, backhaul capacity as well as data forwarding elements (DFE) and SDN controller connections. By invoking massive key performance indicators (KPIs) datasets stemming from a live cellular network endowed with traffic probes, we first introduce a low-complexity slices’ traffics predictor based on a soft gated recurrent unit (GRU). We then build—at each virtual network function—joint multi-slice deep neural networks (DNNs) and train them to estimate the required resources based on the traffic per slice, while not violating two service level agreement (SLA), namely, violation rate -based SLA and resource bounds -based SLA. This is achieved by integrating dataset-dependent generalized non-convex constraints into the DNN offline optimization tasks that are solved via a non-zero sum two-player game strategy. In this respect, we highlight the role of the underlying hyperparameters in the trade-off between overprovisioning and slices’ isolation. Finally, using reliability theory, we provide a closed-form analysis for the lower bound of the so-called reliable convergence probability and showcase the effect of the violation rate on it.

45 citations


Journal ArticleDOI
TL;DR: The comparison analysis among various existing algorithms with TBTS and SLA-LB algorithms show that the proposed methods outperform existing algorithms, even in the scalability situation of the dataset and virtual machines.

45 citations


Journal ArticleDOI
TL;DR: In this article, the authors present a system named Microscaler to automatically identify the scaling-needed services and scale them to meet the Service Level Agreement (SLA) with an optimal cost for microservice applications.
Abstract: Recently, the microservice becomes a popular architecture to construct cloud native systems due to its agility. In cloud native systems, autoscaling is a key enabling technique to adapt to workload changes by acquiring or releasing the right amount of computing resources. However, it becomes a challenging problem in microservice applications, since such an application usually comprises a large number of different microservices with complex interactions. When the performance decreases due to an unpredictable workload peak, it is difficult to pinpoint the scaling-needed services which need to scale out and evaluate how many resources they need. In this paper, we present a novel system named Microscaler to automatically identify the scaling-needed services and scale them to meet the Service Level Agreement (SLA) with an optimal cost for microservice applications. Microscaler first collects the quality of service (QoS) metrics in the service mesh enabled microservice infrastructure. Then, it determines under-provisioning or over-provisioning service instances along the service dependency graph with a novel scaling-needed service criterion named service power. The service dependency graph could be obtained by correlating each request flow in the service mesh. By combining an online learning approach and a step-by-step heuristic approach, Microscaler can precisely reach the optimal service scale meeting the SLA requirements. The experimental evaluations in a microservice benchmark show that Microscaler achieves an average 93% precision in scaling-needed service determination and converges to the optimal service scale faster than several state-of-the-art methods. Moreover, Microscaler is lightweight and flexible enough to work in a large-scale microservice system.

44 citations


Journal ArticleDOI
TL;DR: Fog assisted Computational efficient Wearable sensor networks (FCE-WSN) has been proposed in health monitoring systems for sports athletic using IoT and the experimental results show that the proposed method is user-friendly, reliable, and economical to use regularly.

Journal ArticleDOI
TL;DR: A novel hybrid wavelet time series decomposer and GMDH-ELM ensemble method named Wavelet-GMDH-ELm (WGE) for NFV workload forecasting which predicts and ensembles workload in different time-frequency scales is proposed.

Journal ArticleDOI
TL;DR: A new power-aware VM selection policy has been proposed in this research that helps in VM selection for migration and has been further evaluated using trace-based simulation environment.
Abstract: With the rapid demand for service-oriented computing in association with the growth of cloud computing technologies, large-scale virtualized data centers have been established throughout the globe. These huge data centers consume power at a large scale that results in a high operational cost. The massive carbon footprint from the energy generators is another great issue to deal global warming. It is essential to lower the rate of carbon emission and energy consumption as much as possible. The live-migration-enabled dynamic virtual machine consolidation results in high energy saving. But it also incurs the violation of service level agreement (SLA). Excessive migration may lead to performance degradation and SLA violation. The process of VM selection for migration plays a vital role in the domain of energy-aware cloud computing. Using VM selection policies, VMs are selected for migration. A new power-aware VM selection policy has been proposed in this research that helps in VM selection for migration. The proposed power-aware VM selection policy has been further evaluated using trace-based simulation environment.

Journal ArticleDOI
TL;DR: A genetic algorithm (GA) intelligent latency-aware resource allocation scheme (GI-LARE) is proposed that outperformed the static slicing resource allocation; the spatial branch and bound-based scheme; and, an optimal resource allocation algorithm (ORA) via Monte Carlo simulation.
Abstract: In 5G slice networks, the multi-tenant, multi-tier heterogeneous network will be critical in meeting the quality of service (QoS) requirement of the different slice use cases and in reduction of the capital expenditure (CAPEX) and operational expenditure (OPEX) of mobile network operators. Hence, 5G slice networks should be as flexible as possible to accommodate different network dynamics such as user location and distribution, different slice use case QoS requirements, cell load, intra-cluster interference, delay bound, packet loss probability, and service level agreement (SLA) of mobile virtual network operators (MVNO). Motivated by this condition, this paper addresses a latency-aware dynamic resource allocation problem for 5G slice networks in a multi-tenant, multi-tier heterogeneous environment, for efficient radio resource management. The latency-aware dynamic resource allocation problem is formulated as a maximum utility optimisation problem. The optimisation problem is transformed and the hierarchical decomposition technique is adopted to reduce the complexities in solving the optimisation problem. Furthermore, we propose a genetic algorithm (GA) intelligent latency-aware resource allocation scheme (GI-LARE). We compare GI-LARE with the static slicing (SS) resource allocation; the spatial branch and bound-based scheme; and, an optimal resource allocation algorithm (ORA) via Monte Carlo simulation. Our findings reveal that GI-LARE outperformed these other schemes.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed a model for dynamic reliability-aware service placement based on the simultaneous allocation of the main and backup servers, which aims to minimize the placement cost and maximize the number of admitted services.
Abstract: Network softwarization is one of the major paradigm shifts in the next generation of networks. It enables programmable and flexible management and deployment of the network. Network function virtualization (NFV) is referred to the deployment of software functions running on commodity servers instead of traditional hardware-based middle-boxes. It is an example of network softwarization. In NFV, a service is defined as a chain of software functions named service chain function (SFC). The process of allocating the resources of servers to the services, called service placement, is the most challenging mission in NFV. Dynamic nature of the service arrivals and departures as well as meeting the service level agreement make the service placement problem even more challenging. In this paper, we propose a model for dynamic reliability-aware service placement based on the simultaneous allocation of the main and backup servers. Then, we formulate the dynamic reliability-aware service placement as an infinite horizon Markov decision process (MDP), which aims to minimize the placement cost and maximize the number of admitted services. In the proposed MDP, the number of active services in the network is considered to be the state of the system, and the state of the idle resources is estimated based on it. Also, the number of possible admitted services is considered as the action of the presented MDP. To evaluate each possible action in the proposed MDP, we use a sub-optimal method based on the Viterbi algorithm named Viterbi-based Reliable Static Service Placement (VRSSP) algorithm. We determine the optimal policy based on value iteration method using an algorithm named VRSSP-based Value Iteration (VVI) algorithm. Eventually, through the extensive simulations, the superiority of the proposed model for dynamic reliability-aware service placement compared to the static solutions is inferred.

Journal ArticleDOI
TL;DR: A framework which can show effective performance for achieving the high data center energy efficiency and preventing Service Level Agreement (SLA) violation respectively with the aim of green cloud resources deployment is proposed.

Proceedings ArticleDOI
28 Mar 2020
TL;DR: In this article, a decentralized resource orchestration system named EdgeSlice is proposed for dynamic end-to-end network slicing, which is composed of a performance coordinator and multiple orchestration agents.
Abstract: 5G and edge computing will serve various emerging use cases that have diverse requirements of multiple resources, e.g., radio, transportation, and computing. Network slicing is a promising technology for creating virtual networks that can be customized according to the requirements of different use cases. Provisioning network slices requires end-to-end resource orchestration which is challenging. In this paper, we design a decentralized resource orchestration system named EdgeSlice for dynamic end-to-end network slicing. EdgeSlice introduces a new decentralized deep reinforcement learning (D-DRL) method to efficiently orchestrate end-to-end resources. D-DRL is composed of a performance coordinator and multiple orchestration agents. The performance coordinator manages the resource orchestration policies in all the orchestration agents to ensure the service level agreement (SLA) of network slices. The orchestration agent learns the resource demands of network slices and orchestrates the resource allocation accordingly to optimize the performance of the slices under the constrained networking and computing resources. We design radio, transport and computing manager to enable dynamic configuration of end-to-end resources at runtime. We implement EdgeSlice on a prototype of the end-to-end wireless edge computing network with OpenAirInterface LTE network, OpenDayLight SDN switches, and CUDA GPU platform. The performance of EdgeSlice is evaluated through both experiments and trace-driven simulations. The evaluation results show that EdgeSlice achieves much improvement as compared to baseline in terms of performance, scalability, compatibility.

Proceedings ArticleDOI
15 Apr 2020
TL;DR: Song et al. as discussed by the authors proposed a fine-grained co-location framework for latency-critical workloads with best-effort batch (BE) jobs in datacenters.
Abstract: Cloud service providers improve resource utilization by co-locating latency-critical (LC) workloads with best-effort batch (BE) jobs in datacenters. However, they usually treat an LC workload as a whole when allocating resources to BE jobs and neglect the different features of components of an LC workload. This kind of coarse-grained co-location method leaves a significant room for improvement in resource utilization. Based on the observation of the inconsistent interference tolerance abilities of different LC components, we propose a new abstraction called Servpod, which is a collection of a LC parts that are deployed on the same physical machine together, and show its merits on building a fine-grained co-location framework. The key idea is to differentiate the BE throughput launched with each LC Servpod, i.e., Servpod with high interference tolerance ability can be deployed along with more BE jobs. Based on Servpods, we present Rhythm, a co-location controller that maximizes the resource utilization while guaranteeing LC service's tail latency requirement. It quantifies the interference tolerance ability of each servpod through the analysis of tail-latency contribution. We evaluate Rhythm using LC services in forms of containerized processes and microservices, and find that it can improve the system throughput by 31.7%, CPU utilization by 26.2%, and memory bandwidth utilization by 34% while guaranteeing the SLA (service level agreement).

Journal ArticleDOI
TL;DR: The metrics of peak power efficiency and optimal utilization for heterogeneous physical machines (PMs) are defined and Peak Efficiency Aware Scheduling (PEAS) is proposed, a novel strategy of VM placement and reallocation for achieving dual improvement in performance and energy conservation from the perspective of server clusters.
Abstract: As data centers are consuming massive amount of energy, improving the energy efficiency of cloud computing has emerged as a focus of research. However, it is challenging to reduce energy consumption while maintaining system performance without increasing the risk of Service Level Agreement violations. Most of the existing consolidation approaches for virtual machines (VMs) consider system performance and Quality of Service (QoS) metrics as constraints, which usually results in large scheduling overhead and impossibility to achieve effective improvement in energy efficiency without sacrificing some system performance and cloud service quality. In this paper, we first define the metrics of peak power efficiency and optimal utilization for heterogeneous physical machines (PMs). Then we propose Peak Efficiency Aware Scheduling (PEAS), a novel strategy of VM placement and reallocation for achieving dual improvement in performance and energy conservation from the perspective of server clusters. PEAS allocates and reallocates VMs in an on-line manner and always attempts to maintain PMs working in their peak power efficiency via VM consolidation. Extensive experiments on Cloudsim show that PEAS outperforms several energy-aware consolidation algorithms with regard to energy consumption, system performance as well as multiple QoS metrics.

Journal ArticleDOI
TL;DR: Simulation results confirm that the proposed performance-based SLA (PerSLA) framework for cost, performance, penalties and revenue optimisation is adequate in revenue generation and customers satisfaction.
Abstract: Cost, performance, and penalties are the key factors to revenue generation and customer satisfaction. They have a complex correlation, that gets more complicated when missing a proper framework that unambiguously defines these factors. Service-level agreement (SLA) is the initial document discussing selected parameters as a precondition to business initialisation. The clear definition and application of the SLA is of paramount importance as for modern as a Service online businesses no direct communication between provider and consumer is expected. For the proper implementation of SLA, there should be a satisfactory approach for measuring and monitoring quality of service metrics. This study investigated these issues and proposed performance-based SLA (PerSLA) framework for cost, performance, penalties and revenue optimisation. PerSLA optimises these parameters and maximises both provider revenue and customers satisfaction. Simulation results confirm that the proposed framework is adequate in revenue generation and customers satisfaction. Customers and providers monitor the business with respect to agreed terms and conditions. On violation, the provider is penalised. This agreement increases the trust in relationship between provider and consumer.

Proceedings ArticleDOI
15 Jun 2020
TL;DR: This paper introduces a set of key 5Growth innovations supporting radio slicing, enhanced monitoring and analytics and integration of machine learning, and end-to-end optimization.
Abstract: Spurred by a growing demand for higher-quality mobile services in vertical industries, 5G is integrating a rich set of technologies, traditionally alien to the telco ecosystem, such as machine learning or cloud computing. Despite the initial steps taken in prior research projects in Europe and beyond, additional innovations are needed to support vertical use cases. This is the objective of the 5Growth project: automate vertical support through (i) a portal connecting verticals to 5G platforms (a.k.a. vertical slicer), a multi-domain service orchestrator and a resource management layer, (ii) closed-loop machine-learning-based Service Level Agreement (SLA) control, and (iii) end-to-end optimization. In this paper, we introduce a set of key 5Growth innovations supporting radio slicing, enhanced monitoring and analytics and integration of machine learning.

Journal ArticleDOI
TL;DR: A novel and complete solution for planning network slices of the LTE EPC, tailored for the enhanced Mobile BroadBand use case and shows that the aggregated signaling generation is a Poisson process and the data traffic exhibits self-similarity and long-range-dependence features.
Abstract: 5G is the next telecommunications standards that will enable the sharing of physical infrastructures to provision ultra short-latency applications, mobile broadband services, Internet of Things, etc. Network slicing is the virtualization technique that is expected to achieve that, as it can allow logical networks to run on top of a common physical infrastructure and ensure service level agreement requirements for different services and applications. In this vein, our paper proposes a novel and complete solution for planning network slices of the LTE EPC, tailored for the enhanced Mobile BroadBand use case. The solution defines a framework which consists of: i) an abstraction of the LTE workload generation process, ii) a compound traffic model, iii) performance models of the whole LTE network, and iv) an algorithm to jointly perform the resource dimensioning and network embedding. Our results show that the aggregated signaling generation is a Poisson process and the data traffic exhibits self-similarity and long-range-dependence features. The proposed performance models for the LTE network rely on these results. We formulate the joint optimization problem of resources dimensioning and embedding of a virtualized EPC and propose a heuristic to solve it. By using simulation tools, we validate the proper operation of our solution.

Journal ArticleDOI
TL;DR: A novel marginal resource allocation decision support model to assist cloud providers to manage the cloud SLAs before its execution, covering all possible scenarios, including whether a consumer is new or not, and whether the consumer requests the same or different marginal resources.
Abstract: One of the significant challenges for cloud providers is how to manage resources wisely and how to form a viable service level agreement (SLA) with consumers to avoid any violation or penalties. Some consumers make an agreement for a fixed amount of resources, these being the required resources that are needed to execute its business. Consumers may need additional resources on top of these fixed resources, known as– marginal resources that are only consumed and paid for in case of an increase in business demand. In such contracts, both parties agree on a pricing model in which a consumer pays upfront only for the fixed resources and pays for the marginal resources when they are used. A marginal resource allocation is a challenge for service provider particularly small- to medium-sized service providers as it can affect the usage of their resources and consequently their profits. This paper proposes a novel marginal resource allocation decision support model to assist cloud providers to manage the cloud SLAs before its execution, covering all possible scenarios, including whether a consumer is new or not, and whether the consumer requests the same or different marginal resources. The model relies on the capabilities of the user-based collaborative filtering method with an enhanced top-k nearest neighbor algorithm and a fuzzy logic system to make a decision. The proposed framework assists cloud providers manage their resources in an optimal way and avoid violations or penalties. Finally, the performance of the proposed model is shown through a cloud scenario which demonstrates that our proposed approach can assists cloud providers to manage their resources wisely to avoid violations.

Journal ArticleDOI
TL;DR: An improved multi-verse optimization algorithm for web service composition that is called IMVO algorithm is proposed to improve QoS while satisfying SLA and results show increasing of normalized QoS up to 57% in comparison with the other approaches, especially for service composition problems with SLA.

Journal ArticleDOI
TL;DR: Simulation results demonstrate that the proposed anomaly-based DDoS attack detection framework in cloud environment using a third party auditor (TPA) has following advantages: efficiency, because of the low overhead of computations on CSPs for attack detection; rapid, due to informing a CSP about an attack in a short course of time; and precision, through no false positive detection.
Abstract: Today, the providers of cloud computing services are among the most prominent service suppliers worldwide. Availability of cloud services is one of the most important concerns of cloud service providers (CSPs) and cloud users (CUs). Distributed Denial of Service (DDoS) attacks are common types of security issues which affect cloud services and consequently, can lead to unavailability of the services. Therefore, reducing the effects of DDoS attacks helps CSPs to provide high quality services to CUs. In this paper, first, we propose an anomaly-based DDoS attack detection framework in cloud environment using a third party auditor (TPA). Second, we provide multiple basic assumptions and configurations of cloud environments for establishing simulation tests to evaluate our proposed framework. Then, we provide results of simulation tests to analyze the feasibility of our approach. Simulation results demonstrate that our method for detecting DDoS attacks in CSPs has following advantages: efficiency, because of the low overhead of computations on CSPs for attack detection; rapid, due to informing a CSP about an attack in a short course of time regarding the maximum valid response time which is defined in a service level agreement (SLA); and precision, through no false positive detection as well as a low rate of false negative detection which is < 2% of all scenarios of the simulation tests. Finally, we present a table to compare characteristics of our framework with other ones in the literature.

Journal ArticleDOI
TL;DR: A reputation-based auction mechanism is employed to model the interaction between the business agent who is interested in outsourcing the network coverage and the UAV operators serving in closeby areas, and a permissioned blockchain architecture considering Support Vector Machine (SVM) for real-time autonomous and distributed monitoring of UAV service is proposed.
Abstract: The UAV is emerging as one of the greatest technology developments for rapid network coverage provisioning at affordable cost The aim of this paper is to outsource network coverage of a specific area according to a desired quality of service requirement and to enable various entities in the network to have intelligence to make autonomous decisions using blockchain and auction mechanisms In this regard, by considering a multiple-UAV network where each UAV is associated to its own controlling operator, this paper addresses two major challenges: the selection of the UAV for the desired quality of network coverage and the development of a distributed and autonomous real-time monitoring framework for the enforcement of service level agreement (SLA) For a suitable UAV selection, we employ a reputation-based auction mechanism to model the interaction between the business agent who is interested in outsourcing the network coverage and the UAV operators serving in closeby areas In addition, theoretical analysis is performed to show that the proposed auction mechanism attains a dominant strategy equilibrium For the SLA enforcement and trust model, we propose a permissioned blockchain architecture considering Support Vector Machine (SVM) for real-time autonomous and distributed monitoring of UAV service In particular, smart contract features of the blockchain are invoked for enforcing the SLA terms of payment and penalty, and for quantifying the UAV service reputation Simulation results confirm the accuracy of theoretical analysis and efficacy of the proposed model

Posted Content
TL;DR: This paper designs a decentralized resource orchestration system named EdgeSlice for dynamic end-to-end network slicing and introduces a new decentralized deep reinforcement learning (D-DRL) method to efficiently orchestrate end- to-end resources.
Abstract: 5G and edge computing will serve various emerging use cases that have diverse requirements of multiple resources, e.g., radio, transportation, and computing. Network slicing is a promising technology for creating virtual networks that can be customized according to the requirements of different use cases. Provisioning network slices requires end-to-end resource orchestration which is challenging. In this paper, we design a decentralized resource orchestration system named EdgeSlice for dynamic end-to-end network slicing. EdgeSlice introduces a new decentralized deep reinforcement learning (D-DRL) method to efficiently orchestrate end-to-end resources. D-DRL is composed of a performance coordinator and multiple orchestration agents. The performance coordinator manages the resource orchestration policies in all the orchestration agents to ensure the service level agreement (SLA) of network slices. The orchestration agent learns the resource demands of network slices and orchestrates the resource allocation accordingly to optimize the performance of the slices under the constrained networking and computing resources. We design radio, transport and computing manager to enable dynamic configuration of end-to-end resources at runtime. We implement EdgeSlice on a prototype of the end-to-end wireless edge computing network with OpenAirInterface LTE network, OpenDayLight SDN switches, and CUDA GPU platform. The performance of EdgeSlice is evaluated through both experiments and trace-driven simulations. The evaluation results show that EdgeSlice achieves much improvement as compared to baseline in terms of performance, scalability, compatibility.

Journal ArticleDOI
TL;DR: A container consolidation scheme with usage prediction that jointly exploits the current and predicted CPU utilization based on local history of the considered PMs in the process of the container consolidation to obtain a reliable characterization of overutilized and underutilized PMs.
Abstract: Since service level agreement (SLA) is essentially used to maintain reliable quality of service between cloud providers and clients in cloud environment, there has been a growing effort in reducing power consumption while complying with the SLA by maximizing physical machine (PM)-level utilization and load balancing techniques in infrastructure as a service. However, with the recent introduction of container as a service by cloud providers, containers are increasingly popular and will become the major deployment model in the cloud environment and specifically in platform as a service. Therefore, reducing power consumption while complying with the SLA at virtual machine (VM)-level becomes essential. In this context, we exploit a container consolidation scheme with usage prediction to achieve the above objectives. To obtain a reliable characterization of overutilized and underutilized PMs, our scheme jointly exploits the current and predicted CPU utilization based on local history of the considered PMs in the process of the container consolidation. We demonstrate our solution through simulations on real workloads. The experimental results show that the container consolidation scheme with usage prediction reduces the power consumption, number of container migrations, and average number of active VMs while complying with the SLA.

Journal ArticleDOI
TL;DR: A novel framework that can evaluate and optimize resource allocation strategies effectively and quantitatively and support both the Service Level Agreement (SLA) negotiation and workflow resource allocation optimization efficiently is proposed.
Abstract: Due to the existence of resource variations, it is very challenging for Cloud workflow resource allocation strategies to guarantee a reliable Quality of Service (QoS). Although dozens of resource allocation heuristics have been developed to improve the QoS of Cloud workflow, it is hard to predict their performance under variations because of the lack of accurate modeling and evaluation methods. So far, there is no comprehensive approach that can quantitatively reason the capability of resource allocation strategies or enable the tuning of parameters to optimize resource allocation solutions under variations. To address the above problems, this paper proposes a novel framework that can evaluate and optimize resource allocation strategies effectively and quantitatively. By using the statistical model checker UPPAAL-SMC and supervised learning approaches, our framework can: i) conduct complex QoS queries on resource allocation instances considering resource variations; ii) make quantitative and qualitative comparisons among resource allocation strategies; iii) enable the tuning of parameters to improve the overall QoS; and iv) support the quick optimization of overall workflow QoS under customer requirements and resource variations. The experimental results demonstrate that our automated framework can support both the Service Level Agreement (SLA) negotiation and workflow resource allocation optimization efficiently.

Journal ArticleDOI
TL;DR: A new conceptual SLA model is defined and a multi-dimensional categorization scheme is proposed on its basis to apply the SLA metrics for an in-depth understanding of managing SLAs and the motivation of trends for future research.
Abstract: Recent years have witnessed the booming of big data analytical applications (BDAAs). This trend provides unrivaled opportunities to reveal the latent patterns and correlations embedded in the data, and thus productive decisions may be made. This was previously a grand challenge due to the notoriously high dimensionality and scale of big data, whereas the quality of service offered by providers is the first priority. As BDAAs are routinely deployed on Clouds with great complexities and uncertainties, it is a critical task to manage the service level agreements (SLAs) so that a high quality of service can then be guaranteed. This study performs a systematic literature review of the state of the art of SLA-specific management for Cloud-hosted BDAAs. The review surveys the challenges and contemporary approaches along this direction centering on SLA. A research taxonomy is proposed to formulate the results of the systematic literature review. A new conceptual SLA model is defined and a multi-dimensional categorization scheme is proposed on its basis to apply the SLA metrics for an in-depth understanding of managing SLAs and the motivation of trends for future research.