scispace - formally typeset
Search or ask a question

Showing papers in "IEEE Transactions on Services Computing in 2021"


Journal ArticleDOI
TL;DR: The technique of certificateless signature is utilized to present a new RDPC protocol for checking the integrity of data shared among a group and the security of the scheme is reduced to the assumptions of computational Diffie-Hellman (CDH) and discrete logarithm (DL).
Abstract: Cloud storage service supplies people with an efficient method to share data within a group. The cloud server is not trustworthy, so lots of remote data possession checking (RDPC) protocols are proposed and thought to be an effective way to ensure the data integrity. However, most of RDPC protocols are based on the mechanism of traditional public key infrastructure (PKI), which has obvious security flaw and bears big burden of certificate management. To avoid this shortcoming, identity-based cryptography (IBC) is often chosen to be the basis of RDPC. Unfortunately, IBC has an inherent drawback of key escrow. To solve these problems, we utilize the technique of certificateless signature to present a new RDPC protocol for checking the integrity of data shared among a group. In our scheme, user's private key includes two parts: a partial key generated by the group manager and a secret value chosen by herself/himself. To ensure the right public keys are chosen during the data integrity checking, the public key of each user is associated with her unique identity, for example the name or telephone number. Thus, the certificate is not needed and the problem of key escrow is eliminated too. Meanwhile, the data integrity can still be audited by public verifier without downloading the whole data. In addition, our scheme also supports efficient user revocation from the group. The security of our scheme is reduced to the assumptions of computational Diffie-Hellman (CDH) and discrete logarithm (DL). Experiment results exhibit that the new protocol is very efficient and feasible.

130 citations


Journal ArticleDOI
TL;DR: A privacy-preserving task recommendation scheme (PPTR) for crowdsourcing is proposed, which achieves the task-worker matching while preserving both task privacy and worker privacy.
Abstract: Crowdsourcing is a distributed computing paradigm that utilizes human intelligence or resources from a crowd of workers. Existing solutions of task recommendation in crowdsourcing may leak private and sensitive information about both tasks and workers. To protect privacy, information about tasks and workers should be encrypted before being outsourced to the crowdsourcing platform, which makes the task recommendation a challenging problem. In this paper, we propose a privacy-preserving task recommendation scheme (PPTR) for crowdsourcing, which achieves the task-worker matching while preserving both task privacy and worker privacy. In PPTR, we first exploit the polynomial function to express multiple keywords of task requirements and worker interests. Then, we design a key derivation method based on matrix decomposition, to realize the multi-keyword matching between multiple requesters and multiple workers. Through PPTR, user accountability and user revocation are achieved effectively and efficiently. Extensive privacy analysis and performance evaluation show that PPTR is secure and efficient.

121 citations


Journal ArticleDOI
TL;DR: This study develops an unceRtainty-aware Online Scheduling Algorithm (ROSA) to schedule dynamic and multiple workflows with deadlines that performs better than the five compared algorithms with respect to costs, deviations, deviation, resource utilization, and fairness.
Abstract: Scheduling workflows in cloud service environment has attracted great enthusiasm, and various approaches have been reported up to now. However, these approaches often ignored the uncertainties in the scheduling environment, such as the uncertain task start/execution/finish time, the uncertain data transfer time among tasks, the sudden arrival of new workflows. Ignoring these uncertain factors often leads to the violation of workflow deadlines and increases service renting costs of executing workflows. This study devotes to improving the performance for cloud service platforms by minimizing uncertainty propagation in scheduling workflow applications that have both uncertain task execution time and data transfer time. To be specific, a novel scheduling architecture is designed to control the count of workflow tasks directly waiting on each service instance (e.g., virtual machine and container). Once a task is completed, its start/execution/finish time are available, which means its uncertainties disappearing, and will not affect the subsequent waiting tasks on the same service instance. Thus, controlling the count of waiting tasks on service instances can prohibit the propagation of uncertainties. Based on this architecture, we develop an unce R tainty-aware O nline S cheduling A lgorithm ( ROSA ) to schedule dynamic and multiple workflows with deadlines. The proposed ROSA skillfully integrates both the proactive and reactive strategies. During the execution of the generated baseline schedules, the reactive strategy in ROSA will be dynamically called to produce new proactive baseline schedules for dealing with uncertainties. Then, on the basis of real-world workflow traces, five groups of simulation experiments are carried out to compare ROSA with five typical algorithms. The comparison results reveal that ROSA performs better than the five compared algorithms with respect to costs (up to 56 percent), deviation (up to 70 percent), resource utilization (up to 37 percent), and fairness (up to 37 percent).

116 citations


Journal ArticleDOI
TL;DR: CNMF is proposed, a covering-based quality prediction method for Web services via neighborhood-aware matrix factorization that significantly outperforms eight existing quality prediction methods, including two state-of-the-art methods that also utilize neighborhood information with MF.
Abstract: The number of Web services on the Internet has been growing rapidly. This has made it increasingly difficult for users to find the right services from a large number of functionally equivalent candidate services. Inspecting every Web service for their quality value is impractical because it is very resource consuming. Therefore, the problem of quality prediction for Web services has attracted a lot of attention in the past several years, with a focus on the application of the Matrix Factorization (MF) technique. Recently, researchers have started to employ user similarity to improve MF-based prediction methods for Web services. However, none of the existing methods has properly and systematically addressed two of the major issues: 1) retrieving appropriate neighborhood information, i.e., similar users and services; 2) utilizing full neighborhood information, i.e., both users’ and services’ neighborhood information. In this paper, we propose CNMF, a c overing-based quality prediction method for Web services via n eighborhood-aware m atrix f actorization. The novelty of CNMF is twofold. First, it employs a covering-based clustering method to find similar users and services, which does not require the number of clusters and cluster centroids to be prespecified. Second, it utilizes neighborhood information on both users and services to improve the prediction accuracy. The results of experiments conducted on a real-world dataset containing 1,974,675 Web service invocation records demonstrate that CNMF significantly outperforms eight existing quality prediction methods, including two state-of-the-art methods that also utilize neighborhood information with MF.

114 citations


Journal ArticleDOI
TL;DR: BPay, an outsourcing service fair payment framework based on blockchain in cloud computing is introduced and its security and compatibility analysis indicates that BPay achieves soundness and robust fairness and it is compatible with the Bitcoin blockchain and the Ethereum blockchain.
Abstract: As a milestone in the development of outsourcing services, cloud computing enables an increasing number of individuals and enterprises to enjoy the most advanced services from outsourcing service providers. Because online payment and data security issues are involved in outsourcing services, the mutual distrust between users and service providers may severely impede the wide adoption of cloud computing. Nevertheless, most existing solutions only consider a specific type of services and rely on a trusted third-party to realize fair payment. In this paper, to realize secure and fair payment of outsourcing services in general without relying on any third-party, trusted or not, we introduce BPay, an outsourcing service fair payment framework based on blockchain in cloud computing. We first propose the system architecture, adversary model and design goals of BPay, then describe the design details. Our security and compatibility analysis indicates that BPay achieves soundness and robust fairness and it is compatible with the Bitcoin blockchain and the Ethereum blockchain. The key to the robust fairness and compatibility lies in an all-or-nothing checking-proof protocol and a top-down checking method. In addition, our experimental results show that BPay is computationally efficient. Finally, we present the applications of BPay in outsourcing services.

110 citations


Journal ArticleDOI
TL;DR: This paper goes through the Alexa.com top 4000 most popular sites to identify precisely 500 websites claiming to provide a REST web service API, and analyzes these 500 APIs for key technical features, degree of compliance with REST architectural principles, and for adherence to best practices.
Abstract: Businesses are increasingly deploying their services on the web, in the form of web applications, SOAP services, message-based services, and, more recently, REST services. Although the movement towards REST is widely recognized, there is not much concrete information regarding the technical features being used in the field, such as typical data formats, how HTTP verbs are being used, or typical URI structures, just to name a few. In this paper, we go through the Alexa.com top 4000 most popular sites to identify precisely 500 websites claiming to provide a REST web service API. We analyze these 500 APIs for key technical features, degree of compliance with REST architectural principles (e.g., resource addressability), and for adherence to best practices (e.g., API versioning). We observed several trends (e.g., widespread JSON support, software-generated documentation), but, at the same time, high diversity in services, including differences in adherence to best practices, with only 0.8 percent of services strictly complying with all REST principles. Our results can help practitioners evolve guidelines and standards for designing higher quality services and also understand deficiencies in currently deployed services. Researchers may also benefit from the identification of key research areas, contributing to the deployment of more reliable services.

87 citations


Journal ArticleDOI
TL;DR: Given a service composition and a set of candidate services, Q2C first preprocesses the quality correlations among the candidate services and then constructs a quality correlation index graph to enable efficient queries for quality correlations.
Abstract: As enterprises around the globe embrace globalization, strategic alliances among enterprises have become an important means to gain competitive advantages. Enterprises cooperate to improve the quality or lower the prices of their services, which introduce quality correlations, i.e., the quality of a service is associated with other services. Existing approaches for service composition have not fully and systematically considered the quality correlations between services. In this paper, we propose a novel approach named Q2C ( Q uery of Q uality C orrelation) to systematically model quality correlations and enable efficient queries of quality correlations for service compositions. Given a service composition and a set of candidate services, Q2C first preprocesses the quality correlations among the candidate services and then constructs a quality correlation index graph to enable efficient queries for quality correlations. Extensive experiments are conducted on a real-world web service dataset to demonstrate the effectiveness and efficiency of Q2C.

86 citations


Journal ArticleDOI
TL;DR: A comprehensive survey of the state-of-the-art work on fault tolerance methods proposed for cloud computing is presented and current issues and challenges in cloud fault tolerance are discussed to identify promising areas for future research.
Abstract: This paper presents a comprehensive survey of the state-of-the-art work on fault tolerance methods proposed for cloud computing. The survey classifies fault-tolerance methods into three categories: 1) ReActive Methods (RAMs); 2) PRoactive Methods (PRMs); and 3) ReSilient Methods (RSMs). RAMs allow the system to enter into a fault status and then try to recover the system. PRMs tend to prevent the system from entering a fault status by implementing mechanisms that enable them to avoid errors before they affect the system. On the other hand, recently emerging RSMs aim to minimize the amount of time it takes for a system to recover from a fault. Machine Learning and Artificial Intelligence have played an active role in RSM domain in such a way that the recovery time is mapped to a function to be optimized (i.e., by converging the recovery time to a fraction of milliseconds). As the system learns to deal with new faults, the recovery time will become shorter. In addition, current issues and challenges in cloud fault tolerance are also discussed to identify promising areas for future research.

71 citations


Journal ArticleDOI
TL;DR: A probabilistic matrix factorization approach with implicit correlation regularization to solve the recommendation problem and enhance the recommendation diversity and develops a latent variable model to uncover the latent correlations between APIs by analyzing their co-invocation patterns.
Abstract: Mashups are a dominant approach for building data-centric applications, especially mobile applications, in recent years. Since mashups are predominantly based on public data sources and existing APIs, it requires no sophisticated programming knowledge of people to develop mashup applications. The recent prevalence of open APIs and open data sources in the Big Data era has provided new opportunities for mashup development, but at the same time increase the difficulty of selecting the right services for a given mashup task. The API recommendation for mashup differs from traditional service recommendation tasks in lacking the specific QoS information and formal semantic specification of the APIs, which limits the adoption of many existing methods. Although there are a significant number of service recommendation approaches, most of them focus on improving the recommendation accuracy and work pays attention to the diversity of the recommendation results. Another challenge comes from the existence of both explicit and implicit correlations among the different APIs, which are generally neglected by existing recommendation methods. In this paper, we address the above deficiencies of existing approaches by exploring API recommendation for mashups in the reusable composition context, with the goal of helping developers identify the most appropriate APIs for their composition tasks. In particular, we propose a probabilistic matrix factorization approach with implicit correlation regularization to solve the recommendation problem and enhance the recommendation diversity. We conjecture that the co-invocation of APIs in real-world mashups is driven by both the explicit textual similarity and implicit correlations of APIs such as the similarity or the complementary relationship of APIs. We develop a latent variable model to uncover the latent correlations between APIs by analyzing their co-invocation patterns. We further explore the relationships of topics/categories to the proposed approach. We demonstrate the effectiveness of our approach by conducting extensive experiments on a real dataset crawled from ProgrammableWeb.

69 citations


Journal ArticleDOI
TL;DR: This work believes that the cough sound has the potential to significantly hamper the Covid-19 pandemic across the world and proposes a web tool and underpinning algorithm for the robust, fast, point-of-need identification of the infection.
Abstract: We seek to evaluate the detection performance of a rapid primary screening tool of Covid-19 solely based on the cough sound from 8,380 clinically validated samples with laboratory molecular-test (2,339 Covid-19 positive and 6,041 Covid-19 negative). Samples were clinically labelled according to the results and severity based on quantitative RT-PCR (qRT-PCR) analysis, cycle threshold and lymphocytes count from the patients. Our proposed generic method is a algorithm based on Empirical Mode Decomposition (EMD) with subsequent classification based on a tensor of audio features and deep artificial neural network classifier with convolutional layers called DeepCough'. Two different versions of DeepCough based on the number of tensor dimensions, i.e. DeepCough2D and DeepCough3D, have been investigated. These methods have been deployed in a multi-platform proof-of-concept Web App CoughDetect to administer this test anonymously. Covid-19 recognition results rates achieved a promising AUC (Area Under Curve) of 98.800.83%, sensitivity of 96.431.85%, and specificity of 96.201.74%, and 81.08%5.05% AUC for the recognition of three severity levels. Our proposed web tool and underpinning algorithm for the robust, fast, point-of-need identification of Covid-19 facilitates the rapid detection of the infection. We believe that it has the potential to significantly hamper the Covid-19 pandemic across the world.

69 citations


Journal ArticleDOI
TL;DR: Inspired by the principles of deep learning, a universal deep neural model (DNM) is proposed for making multiple attributes QoS prediction with contexts that achieves superior prediction accuracy in term of mean absolute error (MAE) compared with the state-of-the-art collaborative QoS Prediction techniques.
Abstract: In recent years, various collaborative QoS prediction methods have been put forward to coping with the demand for efficient quality-of-service (QoS) evaluation, by drawing lessons from the recommender systems. However, there still remain some challenging issues on this direction, as how to effectively exploit complex contexts to improve prediction accuracy, and how to realize collaborative QoS prediction of multiple attributes. Inspired by the principles of deep learning, we have proposed a universal deep neural model (DNM) for making multiple attributes QoS prediction with contexts. In this model, contextual features are mapped into a shared latent space to semantically characterize them in the embedding layer. The contextual features with their higher-order interactions are captured through the interaction layer and the perception layers. Multi-tasks prediction is realized by stacking task-specific perception layers on the shared neural layers. Armed with these, DNM provides a powerful framework to integrate with various contextual features to realize multi-attributes QoS prediction. Experimental results from a large-scale QoS-specific dataset demonstrate that DNM achieves superior prediction accuracy in term of mean absolute error (MAE) compared with the state-of-the-art collaborative QoS prediction techniques. Additionally, the DNM model has a good robustness and extensibility on exploiting heterogeneous contextual features.

Journal ArticleDOI
TL;DR: In this paper, the authors present a survey of blockchain security research in three levels, namely, the process level, the data level, and the infrastructure level, which they refer to as the PDI model.
Abstract: Blockchain, an emerging paradigm of secure and shareable computing, is a systematic integration of 1) chain structure for data verification and storage, 2) distributed consensus algorithms for generating and updating data, 3) cryptographic techniques for guaranteeing data transmission and access security, and 4) automated smart contracts for data programming and operations. However, the progress and promotion of Blockchain have been seriously impeded by various security issues in blockchain-based applications. Furthermore, previous research on blockchain security has been mostly technical, overlooking considerable business, organizational, and operational issues. To address this research gap from the perspective of information systems, we review blockchain security research in three levels, namely, the process level, the data level, and the infrastructure level, which we refer to as the PDI model of blockchain security. In this survey study, we first examine the state of blockchain security in the literature. Based on the insights obtained from this initial analysis, we then suggest future directions of research in blockchain security, shedding light on urgent business and industrial concerns in related computing disciplines.

Journal ArticleDOI
TL;DR: A novel routing algorithm, Resource Aware Routing Algorithm (RA-RA), is proposed to solve the DRP-SFC and surpass the performance of other existing algorithms in acceptance rate, throughput, hop count and load balancing.
Abstract: Owing to the Network Function Virtualization (NFV) and Software-Defined Networks (SDN), Service Function Chain (SFC) has become a popular service in SDN and NFV-enabled network. However, as the Virtual Network Function (VNF) of each type is generally multi-instance and flows with SFC requests must traverse a series of specified VNFs in predefined orders, it is a challenge for dynamic SFC formation to optimally select VNF instances and construct paths. Moreover, the load balancing and end-to-end delay need to be paid attention to, when routing flows with SFC requests. Additionally, fine-grained scheduling for traffic at flow level needs differentiated routing which should take flow features into consideration. Unfortunately, traditional algorithms cannot fulfill all these requirements. In this paper, we study the Differentiated Routing Problem considering SFC (DRP-SFC) in SDN and NFV-enabled network. We formulate the DRP-SFC as a Binary Integer Programming (BIP) model aiming to minimize the resource consumption costs of flows with SFC requests. Then a novel routing algorithm, Resource Aware Routing Algorithm (RA-RA), is proposed to solve the DRP-SFC. Performance evaluation shows that RA-RA can efficiently solve the DRP-SFC and surpass the performance of other existing algorithms in acceptance rate, throughput, hop count and load balancing.

Journal ArticleDOI
TL;DR: This paper proposes Multi-level Join VM Placement and Migration algorithms based on the relaxed convex optimization framework to approximate the optimal solution and demonstrates the effectiveness of the proposed algorithms that substantially increases data center efficiency.
Abstract: We study the problem of virtual machine (VM) placement and migration in a data center. In the current approaches, VMs are assigned to physical servers using on-demand provisioning. Such an approach is simple but it often results in a poor performance due to resource fragmentation. Additionally, sub-optimal VM placement usually generates unneeded VM migration and unnecessary cross network traffic. The efficiency of a datacenter therefore significantly depends on how VMs are provisioned and where they are placed. A good placement scheme will not only improve the quality of service but also reduce the operation cost of the data center. In this paper, we study the problem of optimal VM placement and migration to minimize resource usage and power consumption in a data center. We formulate the optimization problem as a joint multiple objective function and solve it by leveraging the framework of convex optimization. Due to the intractable nature of the combinatorial optimization, we then propose Multi-level Join VM Placement and Migration (MJPM) algorithms based on the relaxed convex optimization framework to approximate the optimal solution. The theoretical analysis demonstrates the effectiveness of our proposed algorithms that substantially increases data center efficiency. In addition, our extensive simulation results on different practical topologies show significant performance improvement over the existing approaches.

Journal ArticleDOI
TL;DR: This work proposes a novel method to predict QoS values based on factorization machine, which leverages not only QoS information of users and services but also the user and service neighbor's information and achieves higher prediction accuracy than other QoS prediction methods.
Abstract: With the prevalence of web services, a large number of similar web services are provided by different providers. To select the optimal service among these service candidates, Quality of Service (QoS), representing the non-functional characteristics, plays an important role. To obtain the QoS values of web services, a number of web service QoS prediction methods have been proposed. Collaborative web service QoS prediction is one of the most popular approaches. Based on the historical QoS data, collaborative QoS prediction methods employ memory-based collaborative filtering (CF), model-based CF, or their hybrids to predict QoS values. However, these methods usually only consider the QoS information of similar users and services, neglecting the correlation between them. To enhance the prediction accuracy, we propose a novel method to predict QoS values based on factorization machine, which leverages not only QoS information of users and services but also the user and service neighbor’s information. To evaluate our approach, we conduct experiments on a large-scale real-world dataset with 1,974,675 web service invocations. The experiment results show that our approach achieves higher prediction accuracy than other QoS prediction methods.

Journal ArticleDOI
TL;DR: An integer programming based approach READ-O for solving the robustness-oriented Edge Application Deployment problem as a constrained optimization problem and its NP-hardness is proved, and an approximation algorithm READ-A for efficiently finding near-optimal solutions to large-scale problems is provided.
Abstract: Edge computing (EC) can overcome several limitations of cloud computing. In the EC environment, a service provider can deploy its application instances on edge servers to serve users with low latency. Given a limited budget K for deploying applications in a particular geographical area, some approaches have been proposed to achieves various optimization objectives, e.g., to maximize the servers' coverage, to minimize the average network latency, etc. However, the robustness of the services collectively delivered by the service provider's applications deployed on the edge servers has not been considered at all. This is a critical issue, especially in the highly distributed, dynamic and volatile EC environment. We make the first attempt to tackle this challenge. Specifically, we formulate this Robustness-oriented Edge Application Deployment(READ) problem as a constrained optimization problem and prove its NP-hardness. Then, we provide an integer programming based approach READ-O for solving it precisely, and an approximation algorithm READ-A for efficiently finding near-optimal solutions to large-scale problems. READ-A's approximation ratio is not worse than K/2, which is constant regardless of the total number of edge servers. Evaluation of the widely-used real-world dataset against five representative approaches demonstrates that our approaches can solve the READ problem effectively and efficiently.

Journal ArticleDOI
TL;DR: This paper proposes the first accountable authority revokable CP-ABE based cloud storage system with white-box traceability and auditing, referred to as CryptCloud+, and proves the security and experimental results to demonstrate the utility of the system.
Abstract: Secure cloud storage, which is an emerging cloud service, is designed to protect the confidentiality of outsourced data but also to provide flexible data access for cloud users whose data is out of physical control. Ciphertext-Policy Attribute-Based Encryption (CP-ABE) is regarded as one of the most promising techniques that may be leveraged to secure the guarantee of the service. However, the use of CP-ABE may yield an inevitable security breach which is known as the misuse of access credential (i.e., decryption rights), due to the intrinsic “all-or-nothing” decryption feature of CP-ABE. In this paper, we investigate the two main cases of access credential misuse: one is on the semi-trusted authority side, and the other is on the side of cloud user. To mitigate the misuse, we propose the first accountable authority and revocable CP-ABE based cloud storage system with white-box traceability and auditing, referred to as CryptCloud $^+$ + . We also present the security analysis and further demonstrate the utility of our system via experiments.

Journal ArticleDOI
TL;DR: This paper proposes a QoS-aware RSW allocation algorithm for NCR with joint optimization of latency, energy efficiency, and cost, while considering the characteristics of both RSW and NCR, and develops a heuristic algorithm to obtain a near-optimal solution.
Abstract: Computation offloading for cloud robotics is receiving considerable attention in academic and industrial communities. However, current solutions face challenges: 1) traditional approaches do not consider the characteristics of networked cloud robotics (NCR) (e.g., heterogeneity and robotic cooperation); 2) they fail to capture the characteristics of tasks in a robotic streaming workflow (RSW) (e.g., strict latency requirements and varying task semantics); and 3) they do not consider quality-of-service (QoS) issues for cloud robotics. In this paper, we address these issues by proposing a QoS-aware RSW allocation algorithm for NCR with joint optimization of latency, energy efficiency, and cost, while considering the characteristics of both RSW and NCR. We first propose a novel framework that combines individual robots, robot clusters, and a remote cloud for computation offloading. We then formulate the joint QoS optimization problem for RSW allocation in NCR while considering latency, energy consumption, and operating cost, and show that the problem is NP-hard. Next, we construct a data flow graph based on the characteristics of RSW and NCR, and transform the RSW allocation problem into a mixed-integer linear programming problem. To obtain a near-optimal solution in reasonable time, we also develop a heuristic algorithm. Experiments comparing our approach with others demonstrate significant performance gains, with improved QoS and reduced execution times.

Journal ArticleDOI
TL;DR: A similarity-maintaining privacy preservation (SPP) strategy is designed, which aims to protect the user’s privacy and maintain the utility of user data in the meanwhile, and a location-aware low-rank matrix factorization (LLMF) algorithm is proposed.
Abstract: Web service recommendation plays an important role in building service-oriented systems. QoS-based Web service recommendation has recently gained much attention for providing a promising way to help users find high-quality services. To accurately predict the QoS values of candidate Web services, Web service recommendation systems usually need to collect historical QoS data from users, which will potentially pose a threat to the user’s privacy. However, how to simultaneously protect user’s privacy and make an accurate prediction has not been well studied. By taking these two aspects into consideration, we propose a novel QoS prediction approach for Web service recommendation in this paper. Specifically, we first design a similarity-maintaining privacy preservation (SPP) strategy, which aims to protect the user’s privacy and maintain the utility of user data in the meanwhile. Then, we propose a location-aware low-rank matrix factorization (LLMF) algorithm, which employs the $L_1$ L 1 -norm low-rank matrix factorization to improve the model’s robustness, and combines the matrix factorization model with two kinds of location information (continent, longitude and latitude) in the prediction process. Experimental results on two publicly available real-world Web service QoS datasets demonstrate the effectiveness of our privacy-preserving QoS prediction approach.

Journal ArticleDOI
TL;DR: This paper proposes a novel privacy-preserving deep learning model, namely PDLM, to apply deep learning over the encrypted data under multiple keys and proves that it can achieve users’ privacy preservation and analyzes the efficiency of PDLM in theory.
Abstract: Deep learning has aroused a lot of attention and has been used successfully in many domains, such as accurate image recognition and medical diagnosis. Generally, the training of models requires large, representative datasets, which may be collected from a large number of users and contain sensitive information (e.g., users’ photos and medical information). The collected data would be stored and computed by service providers (SPs) or delegated to an untrusted cloud. The users can neither control how it will be used, nor realize what will be learned from it, which make the privacy issues prominent and severe. To solve the privacy issues, one of the most popular approaches is to encrypt users’ data with their public keys. However, this technique inevitably leads to another challenge that how to train the model based on multi-key encrypted data. In this paper, we propose a novel privacy-preserving deep learning model, namely PDLM, to apply deep learning over the encrypted data under multiple keys. In PDLM, lots of users contribute their encrypted data to SP to learn a specific model. We adopt an effective privacy-preserving calculation toolkit to achieve the training process based on stochastic gradient descent (SGD) in a privacy-preserving manner. We also prove that our PDLM can achieve users’ privacy preservation and analyze the efficiency of PDLM in theory. Finally, we conduct an experiment to evaluate PDLM over two real-world datasets and empirical results demonstrate that our PDLM can effectively and efficiently train the model in a privacy-preserving way.

Journal ArticleDOI
TL;DR: In this article, the authors present a real world case study in order to demonstrate how scalability is positively affected by re-implementing a monolithic architecture (MA) into a microservices architecture (MSA).
Abstract: An increasing interest is growing around the idea of microservices and the promise of improving scalability when compared to monolithic systems. Several companies are evaluating pros and cons of a complex migration. In particular, financial institutions are positioned in a difficult situation due to the economic climate and the appearance of agile competitors that can navigate in a more flexible legal framework and started their business since day one with more agile architectures and without being bounded to outdated technological standard. In this paper, we present a real world case study in order to demonstrate how scalability is positively affected by re-implementing a monolithic architecture (MA) into a microservices architecture (MSA). The case study is based on the FX Core system, a mission critical system of Danske Bank, the largest bank in Denmark and one of the leading financial institutions in Northern Europe. The technical problem that has been addressed and solved in this paper is the identification of a repeatable migration process that can be used to convert a real world Monolithic architecture into a Microservices architecture in the specific setting of financial domain, typically characterized by legacy systems and batch-based processing on heterogeneous data sources.

Journal ArticleDOI
TL;DR: A return to roots is proposed by defining a Model-Driven Engineering (MDE) methodology that supports automation of BDA based on model specification that lets customers declare requirements to be achieved by an abstract Big Data platform and smart engines deploy the Big Data pipeline carrying out the analytics on a specific instance of such platform.
Abstract: The Big Data revolution promises to build a data-driven ecosystem where better decisions are supported by enhanced analytics and data management. However, major hurdles still need to be overcome on the road that leads to commoditization and wide adoption of Big Data Analytics (BDA). Big Data complexity is the first factor hampering the full potential of BDA. The opacity and variety of Big Data technologies and computations, in fact, make BDA a failure prone and resource-intensive process, which requires a trial-and-error approach. This problem is even exacerbated by the fact that current solutions to Big Data application development take a bottom-up approach, where the last technology release drives application development. Selection of the best Big Data platform, as well as of the best pipeline to execute analytics, represents then a deal breaker. In this paper, we propose a return to roots by defining a Model-Driven Engineering (MDE) methodology that supports automation of BDA based on model specification. Our approach lets customers declare requirements to be achieved by an abstract Big Data platform and smart engines deploy the Big Data pipeline carrying out the analytics on a specific instance of such platform. Driven by customers’ requirements, our methodology is based on an OWL-S ontology of Big Data services and on a compiler transforming OWL-S service compositions in workflows that can be directly executed on the selected platform. The proposal is experimentally evaluated in a real-world scenario focusing on the threat detection system of SAP.

Journal ArticleDOI
TL;DR: A privacy-preserving clinical decision support system using Naïve Bayesian (NB) classifier, hereafter referred to as Peneus, designed for the outsourced cloud computing environment, which achieves the goal of patient health status monitoring without privacy leakage to unauthorized parties.
Abstract: In this paper, we propose a privacy-preserving clinical decision support system using Naive Bayesian (NB) classifier, hereafter referred to as Peneus, designed for the outsourced cloud computing environment. Peneus allows one to use patient health information to train the NB classifier privately, which can then be used to predict a patient's (undiagnosed) disease based on his/her symptoms in a single communication round. Specifically, we design secure Single Instruction Multiple Data (SIMD) integer circuits using the fully homomorphic encryption scheme, which can greatly increase the performance compared with the original secure integer circuit. Then, we present a privacy-preserving historical Personal Health Information (PHI) aggregation protocol to allow different PHI sources to be securely aggregated without the risk of compromising the privacy of individual data owner. Also, secure NB classifier is constructed to achieve secure disease prediction in the cloud without the help of an additional non-colluding computation server. We then demonstrate that Peneus achieves the goal of patient health status monitoring without privacy leakage to unauthorized parties, as well as the utility and the efficiency of Peneus using simulations and analysis.

Journal ArticleDOI
TL;DR: A Location-based Matrix Factorization using a Preference Propagation method (LMF-PP) to address the cold start problem and shows better performance than existing approaches in cold start environments as well as in warm start environments.
Abstract: Many web-based software systems have been developed in the form of composite services. It is important to accurately predict the Quality of Service (QoS) value of atomic web services because the performance of such composite services depends greatly on the performance of the atomic web service adopted. In recent years, collaborative filtering based methods for predicting the web service QoS values have been proposed. However, they are mainly faced with a cold start problem that is difficult to make reliable prediction due to highly sparse historical data, newly introduced users and web services, and the existing work only deals with the case of newly introduced users. In this article, we propose a Location-based Matrix Factorization using a Preference Propagation method (LMF-PP) to address the cold start problem. LMF-PP fuses invocation and neighborhood similarity, and then the fused similarity is utilized by preference propagation. LMF-PP is compared with existing approaches on the real world dataset. Based on the experimental results, LMF-PP shows better performance than existing approaches in cold start environments as well as in warm start environments.

Journal ArticleDOI
TL;DR: This article introduces a lightweight adaptive monitoring framework suitable for smart IoT devices with limited processing capabilities based on a low-cost adaptive and probabilistic learning model capable of capturing at runtime the current evolution and variability of the data stream.
Abstract: Internet-enabled physical devices with “smart” processing capabilities are becoming the tools for understanding the complexity of the global inter-connected world we inhabit. The Internet of Things (IoT) churns tremendous amounts of data flooding from devices scattered across multiple locations to the processing engines of almost all industry sectors. However, as the number of “things” surpasses the population of the technology-enabled world, real-time processing and energy-efficiency are great challenges of the big data era transitioning to IoT. In this article, we introduce a lightweight adaptive monitoring framework suitable for smart IoT devices with limited processing capabilities. Our framework, inexpensively and in place dynamically adjusts the monitoring intensity and the amount of data disseminated through the network based on a low-cost adaptive and probabilistic learning model capable of capturing at runtime the current evolution and variability of the data stream. By accomplishing this, energy consumption and data volume are reduced, allowing IoT devices to preserve battery and ease processing on cloud computing and streaming services. Experiments on real-world data from cloud services, internet security services, wearables and intelligent transportation services, show that our framework achieves a balance between efficiency and accuracy. Specifically, our framework reduces data volume by 74 percent, energy consumption by at least 71 percent, while maintaining accuracy always above 89 percent.

Journal ArticleDOI
TL;DR: A green service composition approach that gives priority to those composite services that are hosted on the same virtual machine, physical server, or edge switch with end-to-end QoS guarantee and fulfills the green service compositions optimization by minimizing the energy and network resource consumption on physical servers and switches in cloud data centers.
Abstract: With the increasing popularity of cloud computing, many notable quality of service (QoS)-aware service composition approaches have been incorporated in service-oriented cloud computing systems However, these approaches are implemented without considering the energy and network resource consumption of the composite services The increases in energy and network resource consumption resulting from these compositions can incur a high cost in data centers In this paper, the trade-off among QoS performance, energy consumption, and network resource consumption in a service composition process is first analyzed Then, a green service composition approach is proposed It gives priority to those composite services that are hosted on the same virtual machine, physical server, or edge switch with end-to-end QoS guarantee It fulfills the green service composition optimization by minimizing the energy and network resource consumption on physical servers and switches in cloud data centers Experimental results indicate that, with comparisons to other approaches, our approach saves 20-50 percent of energy consumption and 10-50 percent of network resource consumption

Journal ArticleDOI
TL;DR: This work proposes a temporal task scheduling algorithm investigating the temporal variation in green hybrid cloud to schedule all tasks within their delay constraints, and explicitly presents a mathematical equation of service rates and task refusal.
Abstract: A growing number of global companies select Green Data Centers (GDCs) to manage their delay-constrained applications. The fast growth of users’ tasks dramatically increases the energy consumed by GDCs owned by a company, e.g., Google and Amazon. The random nature of tasks brings a big challenge of scheduling tasks of each application with limited infrastructure resources of GDCs. Therefore, hybrid cloud is widely employed to smartly outsource some tasks to public clouds. However, the temporal variation in many factors including revenue, price of power grid, solar irradiance, wind speed, price of public clouds makes it challenging to schedule all tasks of each application in a cost-effective way while strictly meeting their expected delay constraints. This work proposes a temporal task scheduling algorithm investigating the temporal variation in green hybrid cloud to schedule all tasks within their delay constraints. Besides, it explicitly presents a mathematical equation of service rates and task refusal. The maximization problem is formulated and tackled by the proposed hybrid optimization algorithm called Genetic Simulated-annealing-based particle swarm optimization. Trace-driven experiments demonstrate that larger profit are achieved than several existing scheduling algorithms.

Journal ArticleDOI
TL;DR: In this article, the authors proposed a generalized Nesterov's accelerated gradient (NAG) method for non-negative latent factor (NLF) model with a single latent factor-dependent, nonnegative and multiplicative update (SLF-NMU) algorithm.
Abstract: A non-negative latent factor (NLF) model with a single latent factor-dependent, non-negative and multiplicative update (SLF-NMU) algorithm is frequently adopted to extract useful knowledge from non-negative data represented by high-dimensional and sparse (HiDS) matrices arising from various service applications. However, its convergence rate is rather slow. To address this issue, this study proposes a Generalized Nesterov's acceleration-incorporated, Non-negative and Adaptive Latent Factor (GNALF) model. It results from a) incorporating a generalized Nesterov's accelerated gradient (NAG) method into an SLF-NMU algorithm, thereby achieving an NAG-incorporated and element-oriented non-negative (NEN) algorithm to perform efficient parameter update; and b) making its regularization and acceleration hyper-parameters self-adaptive via incorporating the principle of a particle swarm optimization algorithm into the training process, thereby implementing a highly adaptive and practical model. Empirical studies on six large sparse matrices from different recommendation service applications show that a GNALF model achieves very high convergence rate without the need of hyper-parameter tuning, making its computational efficiency significantly higher than state-of-the-art models. Meanwhile, such efficiency gain does not result in accuracy loss, since its prediction accuracy is comparable with its peers. Hence, it can better serve practical service applications with real-time demands.

Journal ArticleDOI
TL;DR: This work designs a distributed auction mechanism to fairly allocate the tasks, and determines the trading prices of the resources, and proves that the proposed auction mechanism can achieve certain desirable properties, such as computational efficiency, individual rationality, truthfulness guarantee of the bidders, and budget balance.
Abstract: In mobile cloud computing, offloading resource-demanded applications from mobile devices to remote cloud servers can alleviate the resource scarcity of mobile devices, whereas long distance communication may incur high communication latency and energy consumption. As an alternative, fortunately, recent studies show that exploiting the unused resources of the nearby mobile devices for task execution can reduce the energy consumption and communication latency. Nevertheless, it is non-trivial to encourage mobile devices to share their resources or execute tasks for others. To address this issue, we construct an auction model to facilitate the resource trading between the owner of the tasks and the mobile devices participating in task execution. Specifically, the owners of the tasks act as bidders by submitting bids to compete for the resources available at mobile devices. We design a distributed auction mechanism to fairly allocate the tasks, and determine the trading prices of the resources. Moreover, an efficient payment evaluation process is proposed to prevent against the possible dishonest activity of the seller on the payment decision, through the collaboration of the buyers. We prove that the proposed auction mechanism can achieve certain desirable properties, such as computational efficiency, individual rationality, truthfulness guarantee of the bidders, and budget balance. Simulation results validate the performance of the proposed auction mechanism.

Journal ArticleDOI
TL;DR: This paper proposes and compares different configuration models for deploying a containerized software system, Inspired by Google Kubernetes, and develops novel non-state-space and state-space analytic models for container availability analysis.
Abstract: Operating system (OS) containers enabling the microservice-oriented architecture are becoming popular in the context of Cloud services. Containers provide the ability to create lightweight and portable runtime environments that decouple the application requirements from the characteristics of the underlying system. Services built on containers have a small resource footprint in terms of processing, storage, memory and network, allowing a more dense deployment environment. While the performance of such containers is addressed in few previous studies, understanding the failure-repair behavior of the containers remains unexplored. In this paper, from an availability point of view, we propose and compare different configuration models for deploying a containerized software system. Inspired by Google Kubernetes, a container management system, these configurations are characterized with a failure response and migration service. We develop novel non-state-space (i.e., fault tree) and state-space (i.e., stochastic reward net) analytic models for container availability analysis. Analytical as well as simulative solutions are obtained for the developed models. Our analysis provides insights on k out-of N availability and sensitivity of system availability for key system parameters. Finally, we build an open-source software tool powered by these models. The tool helps a Cloud administrator to assess the availability of a containerized system and to conduct a what-if analysis based on user-provided parameters and configurations.