scispace - formally typeset
Search or ask a question

Showing papers on "Server published in 2012"


Proceedings ArticleDOI
03 Mar 2012
TL;DR: This work identifies the key micro-architectural needs of scale-out workloads, calling for a change in the trajectory of server processors that would lead to improved computational density and power efficiency in data centers.
Abstract: Emerging scale-out workloads require extensive amounts of computational resources. However, data centers using modern server hardware face physical constraints in space and power, limiting further expansion and calling for improvements in the computational density per server and in the per-operation energy. Continuing to improve the computational resources of the cloud while staying within physical constraints mandates optimizing server efficiency to ensure that server hardware closely matches the needs of scale-out workloads.In this work, we introduce CloudSuite, a benchmark suite of emerging scale-out workloads. We use performance counters on modern servers to study scale-out workloads, finding that today's predominant processor micro-architecture is inefficient for running these workloads. We find that inefficiency comes from the mismatch between the workload needs and modern processors, particularly in the organization of instruction and data memory systems and the processor core micro-architecture. Moreover, while today's predominant micro-architecture is inefficient when executing scale-out workloads, we find that continuing the current trends will further exacerbate the inefficiency in the future. In this work, we identify the key micro-architectural needs of scale-out workloads, calling for a change in the trajectory of server processors that would lead to improved computational density and power efficiency in data centers.

860 citations


Journal ArticleDOI
Cong Wang1, Qian Wang1, Kui Ren1, Ning Cao, Wenjing Lou 
TL;DR: This paper proposes a flexible distributed storage integrity auditing mechanism, utilizing the homomorphic token and distributed erasure-coded data, which is highly efficient and resilient against Byzantine failure, malicious data modification attack, and even server colluding attacks.
Abstract: Cloud storage enables users to remotely store their data and enjoy the on-demand high quality cloud applications without the burden of local hardware and software management. Though the benefits are clear, such a service is also relinquishing users' physical possession of their outsourced data, which inevitably poses new security risks toward the correctness of the data in cloud. In order to address this new problem and further achieve a secure and dependable cloud storage service, we propose in this paper a flexible distributed storage integrity auditing mechanism, utilizing the homomorphic token and distributed erasure-coded data. The proposed design allows users to audit the cloud storage with very lightweight communication and computation cost. The auditing result not only ensures strong cloud storage correctness guarantee, but also simultaneously achieves fast data error localization, i.e., the identification of misbehaving server. Considering the cloud data are dynamic in nature, the proposed design further supports secure and efficient dynamic operations on outsourced data, including block modification, deletion, and append. Analysis shows the proposed scheme is highly efficient and resilient against Byzantine failure, malicious data modification attack, and even server colluding attacks.

678 citations


Patent
17 Aug 2012
TL;DR: In this article, a collaborative web browsing system has been proposed for computerized social networks and e-commerce and facilitating ad-hoc screen sharing and co-browsing between users of a social network.
Abstract: The present invention is directed towards to computerized social networks and e-commerce and facilitating ad-hoc screen sharing and co-browsing between users of a social network The collaborative web browsing system has document object model element interaction detection The collaborative browsing method comprises a server computer having a Shopping With A Friend (SWAF) server engine coupled to a database, a SWAF client engine coupled to the SWAF server engine and a plurality of client computers each having a web browser program that runs the SWAF client engine The web browser program does not include a collaboration plug-in

642 citations


Patent
17 Aug 2012
TL;DR: In this paper, the authors proposed a collaborative web browsing method comprising providing a server computer having a Shopping With A Friend (SWAF) server engine coupled with a database, a SWAF client engine coupled to the SWAF server engine and a plurality of client computers each having a web browser program that runs the SCA client engine.
Abstract: The present invention is directed towards computerized social networks and e-commerce including facilitating ad-hoc screen sharing and co-browsing between users of a social network. The collaborative web browsing method comprising providing a server computer having a Shopping With A Friend (SWAF) server engine coupled to a database, a SWAF client engine coupled to the SWAF server engine and a plurality of client computers each having a web browser program that runs the SWAF client engine. The web browser program does not include a collaboration plug-in.

635 citations


Patent
20 Aug 2012
TL;DR: In this article, the authors proposed a collaborative browsing method for computerized social networks and e-commerce and facilitating ad-hoc screen sharing and co-browsing between users of a social network.
Abstract: The present invention is directed towards to computerized social networks and e-commerce and facilitating ad-hoc screen sharing and co-browsing between users of a social network. The collaborative browsing is integrated with social networks. The collaborative browsing method comprises a server computer having a Shopping With A Friend (SWAF) server engine coupled to a database, a SWAF client engine coupled to the SWAF server engine and a plurality of client computers each having a web browser program that runs the SWAF client engine. The web browser program does not include a collaboration plug-in.

634 citations


Journal ArticleDOI
TL;DR: A simulation environment for energy-aware cloud computing data centers is presented and the effectiveness of the simulator in utilizing power management schema, such as voltage scaling, frequency scaling, and dynamic shutdown that are applied to the computing and networking components are demonstrated.
Abstract: Cloud computing data centers are becoming increasingly popular for the provisioning of computing resources. The cost and operating expenses of data centers have skyrocketed with the increase in computing capacity. Several governmental, industrial, and academic surveys indicate that the energy utilized by computing and communication units within a data center contributes to a considerable slice of the data center operational costs. In this paper, we present a simulation environment for energy-aware cloud computing data centers. Along with the workload distribution, the simulator is designed to capture details of the energy consumed by data center components (servers, switches, and links) as well as packet-level communication patterns in realistic setups. The simulation results obtained for two-tier, three-tier, and three-tier high-speed data center architectures demonstrate the effectiveness of the simulator in utilizing power management schema, such as voltage scaling, frequency scaling, and dynamic shutdown that are applied to the computing and networking components.

599 citations


Journal ArticleDOI
TL;DR: The GalaxyWEB server predicts protein structure from sequence by template-based modeling and refines loop or terminus regions by ab initio modeling and generates reliable core structures from multiple templates and re-builds unreliable loops or termini by using an optimization-based refinement method.
Abstract: Three-dimensional protein structures provide invaluable information for understanding and regulating biological functions of proteins. The GalaxyWEB server predicts protein structure from sequence by template-based modeling and refines loop or terminus regions by ab initio modeling. This web server is based on the method tested in CASP9 (9th Critical Assessment of techniques for protein Structure Prediction) as ‘Seok-server’, which was assessed to be among top performing template-based modeling servers. The method generates reliable core structures from multiple templates and re-builds unreliable loops or termini by using an optimization-based refinement method. In addition to structure prediction, a user can also submit a refinement only job by providing a starting model structure and locations of loops or termini to refine. The web server can be freely accessed at http://galaxy.seoklab.org/.

535 citations


Journal ArticleDOI
TL;DR: A dynamic offloading algorithm based on Lyapunov optimization is presented, which has low complexity to solve the offloading problem and shows that the proposed algorithm saves more energy than the existing algorithm while meeting the requirement of application execution time.
Abstract: Offloading is an effective method for extending the lifetime of handheld mobile devices by executing some components of applications remotely (e.g., on the server in a data center or in a cloud). In this article, to achieve energy saving while satisfying given application execution time requirement, we present a dynamic offloading algorithm, which is based on Lyapunov optimization. The algorithm has low complexity to solve the offloading problem (i.e., to determine which software components to execute remotely given available wireless network connectivity). Performance evaluation shows that the proposed algorithm saves more energy than the existing algorithm while meeting the requirement of application execution time.

522 citations


Patent
20 Feb 2012
TL;DR: In this paper, the authors present a request identification and parsing process to locate object metadata and to handle the request in accordance therewith, where different types of metadata exist for a particular object, where metadata in a configuration file is overridden by metadata in response header or request string, with metadata in the request string taking precedence.
Abstract: To serve content through a content delivery network (CDN), the CDN must have some information about the identity, characteristics and state of its target objects. Such additional information is provided in the form of object metadata, which according to the invention can be located in the request string itself, in the response headers from the origin server, in a metadata configuration file distributed to CDN servers, or in a per-customer metadata configuration file. CDN content servers execute a request identification and parsing process to locate object metadata and to handle the request in accordance therewith. Where different types of metadata exist for a particular object, metadata in a configuration file is overridden by metadata in a response header or request string, with metadata in the request string taking precedence.

504 citations


Journal ArticleDOI
TL;DR: This paper addresses the construction of an efficient PDP scheme for distributed cloud storage to support the scalability of service and data migration, in which it considers the existence of multiple cloud service providers to cooperatively store and maintain the clients' data.
Abstract: Provable data possession (PDP) is a technique for ensuring the integrity of data in storage outsourcing. In this paper, we address the construction of an efficient PDP scheme for distributed cloud storage to support the scalability of service and data migration, in which we consider the existence of multiple cloud service providers to cooperatively store and maintain the clients' data. We present a cooperative PDP (CPDP) scheme based on homomorphic verifiable response and hash index hierarchy. We prove the security of our scheme based on multiprover zero-knowledge proof system, which can satisfy completeness, knowledge soundness, and zero-knowledge properties. In addition, we articulate performance optimization mechanisms for our scheme, and in particular present an efficient method for selecting optimal parameter values to minimize the computation costs of clients and storage service providers. Our experiments show that our solution introduces lower computation and communication overheads in comparison with noncooperative approaches.

473 citations


Proceedings ArticleDOI
01 Jul 2012
TL;DR: The preliminary simulation results show that optimal task partitioning algorithms significantly affect response time with heterogeneous latencies and server compute powers, and high-powered cloudlets are technically feasible and indeed help reduce overall processing time when face recognition applications run on mobile devices using the cloud as the backend servers.
Abstract: Face recognition applications for airport security and surveillance can benefit from the collaborative coupling of mobile and cloud computing as they become widely available today. This paper discusses our work with the design and implementation of face recognition applications using our mobile-cloudlet-cloud architecture named MOCHA and its initial performance results. The challenge lies with how to perform task partitioning from mobile devices to cloud and distribute compute load among cloud servers (cloudlet) to minimize the response time given diverse communication latencies and server compute powers. Our preliminary simulation results show that optimal task partitioning algorithms significantly affect response time with heterogeneous latencies and compute powers. Motivated by these results, we design, implement, and validate the basic functionalities of MOCHA as a proof-of-concept, and develop algorithms that minimize the overall response time for face recognition. Our experimental results demonstrate that high-powered cloudlets are technically feasible and indeed help reduce overall processing time when face recognition applications run on mobile devices using the cloud as the backend servers.

Proceedings Article
28 May 2012
TL;DR: Since there are multiple owners (patients) in a PHR system and every owner would encrypt her PHR files using a different set of cryptographic keys, it is important to reduce the key distribution complexity in such multi-owner settings.
Abstract: Online personal health record (PHR) enables patients to manage their own medical records in a centralized way, which greatly facilitates the storage, access and sharing of personal health data. With the emergence of cloud computing, it is attractive for the PHR service providers to shift their PHR applications and storage into the cloud, in order to enjoy the elastic resources and reduce the operational cost. However, by storing PHRs in the cloud, the patients lose physical control to their personal health data, which makes it necessary for each patient to encrypt her PHR data before uploading to the cloud servers. Under encryption, it is challenging to achieve fine-grained access control to PHR data in a scalable and efficient way. For each patient, the PHR data should be encrypted so that it is scalable with the number of users having access. Also, since there are multiple owners (patients) in a PHR system and every owner would encrypt her PHR files using a different set of cryptographic keys, it is important to reduce the key distribution complexity in such multi-owner settings. Existing cryptographic enforced access control schemes are mostly designed for the single-owner scenarios.

Journal ArticleDOI
TL;DR: A novel approximate analytical model is described that allows cloud operators to determine the relationship between the number of servers and input buffer size and the performance indicators such as mean number of tasks in the system, blocking probability, and probability that a task will obtain immediate service.
Abstract: Successful development of cloud computing paradigm necessitates accurate performance evaluation of cloud data centers. As exact modeling of cloud centers is not feasible due to the nature of cloud centers and diversity of user requests, we describe a novel approximate analytical model for performance evaluation of cloud server farms and solve it to obtain accurate estimation of the complete probability distribution of the request response time and other important performance indicators. The model allows cloud operators to determine the relationship between the number of servers and input buffer size, on one side, and the performance indicators such as mean number of tasks in the system, blocking probability, and probability that a task will obtain immediate service, on the other.

Proceedings ArticleDOI
14 Nov 2012
TL;DR: This work measures three popular video streaming services -- Hulu, Netflix, and Vudu -- and finds that accurate client-side bandwidth estimation above the HTTP layer is hard, and rate selection based on inaccurate estimates can trigger a feedback loop, leading to undesirably variable and low-quality video.
Abstract: Today's commercial video streaming services use dynamic rate selection to provide a high-quality user experience. Most services host content on standard HTTP servers in CDNs, so rate selection must occur at the client. We measure three popular video streaming services -- Hulu, Netflix, and Vudu -- and find that accurate client-side bandwidth estimation above the HTTP layer is hard. As a result, rate selection based on inaccurate estimates can trigger a feedback loop, leading to undesirably variable and low-quality video. We call this phenomenon the "downward spiral effect", and we measure it on all three services, present insights into its root causes, and validate initial solutions to prevent it.

Proceedings ArticleDOI
25 Mar 2012
TL;DR: An efficient 2-approximation algorithm for the optimal selection of data centers in the distributed cloud and a heuristic for partitioning the requested resources for the task amongst the chosen data centers and racks are developed.
Abstract: We consider resource allocation algorithms for distributed cloud systems, which deploy cloud-computing resources that are geographically distributed over a large number of locations in a wide-area network. This distribution of cloud-computing resources over many locations in the network may be done for several reasons, such as to locate resources closer to users, to reduce bandwidth costs, to increase availability, etc. To get the maximum benefit from a distributed cloud system, we need efficient algorithms for resource allocation which minimize communication costs and latency. In this paper, we develop efficient resource allocation algorithms for use in distributed clouds. Our contributions are as follows: Assuming that users specify their resource needs, such as the number of virtual machines needed for a large computational task, we develop an efficient 2-approximation algorithm for the optimal selection of data centers in the distributed cloud. Our objective is to minimize the maximum distance, or latency, between the selected data centers. Next, we consider use of a similar algorithm to select, within each data center, the racks and servers where the requested virtual machines for the task will be located. Since the network inside a data center is structured and typically a tree, we make use of this structure to develop an optimal algorithm for rack and server selection. Finally, we develop a heuristic for partitioning the requested resources for the task amongst the chosen data centers and racks. We use simulations to evaluate the performance of our algorithms over example distributed cloud systems and find that our algorithms provide significant gains over other simpler allocation algorithms.

Proceedings ArticleDOI
25 Mar 2012
TL;DR: This work shows that the widely-used Best-Fit scheduling algorithm is not throughput-optimal, and presents alternatives which achieve any arbitrary fraction of the capacity region of the cloud, and studies the delay performance of these alternative algorithms through simulations.
Abstract: Cloud computing services are becoming ubiquitous, and are starting to serve as the primary source of computing power for both enterprises and personal computing applications. We consider a stochastic model of a cloud computing cluster, where jobs arrive according to a stochastic process and request virtual machines (VMs), which are specified in terms of resources such as CPU, memory and storage space. While there are many design issues associated with such systems, here we focus only on resource allocation problems, such as the design of algorithms for load balancing among servers, and algorithms for scheduling VM configurations. Given our model of a cloud, we first define its capacity, i.e., the maximum rates at which jobs can be processed in such a system. Then, we show that the widely-used Best-Fit scheduling algorithm is not throughput-optimal, and present alternatives which achieve any arbitrary fraction of the capacity region of the cloud. We then study the delay performance of these alternative algorithms through simulations.

Patent
12 Dec 2012
TL;DR: A method, apparatus, computer readable medium, computer system, wireless or wired network, or system to provide an online and/or mobile security of a user's privacy and privacy method of internet or mobile access or system using encryption technologies and or filters to access data, encrypt and decrypt data, sync data, secure data storage and process data using cloud technology across many different networks and fiber optic communications from an endpoint accessed through multiple devices, browsers, operating systems, networks, servers, storage, software, applications or services integrated in a public cloud or a private cloud within an enterprise,
Abstract: A method, apparatus, computer readable medium, computer system, wireless or wired network, or system to provide an online and/or mobile security of a user's privacy and/or security method of internet or mobile access or system, apparatus, computer readable medium, or system using encryption technologies and/or filters to access data, encrypt and/or decrypt data, sync data, secure data storage and/or process data using cloud technology across many different networks and/or fiber optic communications from an endpoint accessed through multiple devices, browsers, operating systems, networks, servers, storage, software, applications or services integrated in a public cloud or a private cloud within an enterprise, a social network, big data analytics or electronic surveillance tracking or some mashup of two or more to prevent the unauthorized collecting, tracking and/or analysis of a user's personal data by a third party and/or for generating relevant advertising, mobile, internet social messaging, internet posted promotions or offers for products and/or services.

Proceedings ArticleDOI
22 Aug 2012
TL;DR: A machine learning-based system for the detection of malware on Android devices that extracts a number of features and trains a One-Class Support Vector Machine in an offline (off-device) manner, in order to leverage the higher computing power of a server or cluster of servers.
Abstract: With the recent emergence of mobile platforms capable of executing increasingly complex software and the rising ubiquity of using mobile platforms in sensitive applications such as banking, there is a rising danger associated with malware targeted at mobile devices. The problem of detecting such malware presents unique challenges due to the limited resources avalible and limited privileges granted to the user, but also presents unique opportunity in the required metadata attached to each application. In this article, we present a machine learning-based system for the detection of malware on Android devices. Our system extracts a number of features and trains a One-Class Support Vector Machine in an offline (off-device) manner, in order to leverage the higher computing power of a server or cluster of servers.

01 Aug 2012
TL;DR: This document improves on that situation by enabling the administrators of domain names to specify the keys used in that domain's TLS servers, which requires matching improvements in TLS client software, but no change in TLS server software.
Abstract: Encrypted communication on the Internet often uses Transport Layer Security (TLS), which depends on third parties to certify the keys used. This document improves on that situation by enabling the administrators of domain names to specify the keys used in that domain's TLS servers. This requires matching improvements in TLS client software, but no change in TLS server software. [STANDARDS- TRACK]

Proceedings ArticleDOI
04 Jun 2012
TL;DR: This paper studies online algorithms for determining the number of servers to leave on in each data center, and then uses these algorithms to study the environmental potential of geographical load balancing (GLB) and to understand what portfolio of renewable energy is most effective.
Abstract: It has recently been proposed that Internet energy costs, both monetary and environmental, can be reduced by exploiting temporal variations and shifting processing to data centers located in regions where energy currently has low cost. Lightly loaded data centers can then turn off surplus servers. This paper studies online algorithms for determining the number of servers to leave on in each data center, and then uses these algorithms to study the environmental potential of geographical load balancing (GLB). A commonly suggested algorithm for this setting is “receding horizon control” (RHC), which computes the provisioning for the current time by optimizing over a window of predicted future loads. We show that RHC performs well in a homogeneous setting, in which all servers can serve all jobs equally well; however, we also prove that differences in propagation delays, servers, and electricity prices can cause RHC perform badly, So, we introduce variants of RHC that are guaranteed to perform as well in the face of such heterogeneity. These algorithms are then used to study the feasibility of powering a continent-wide set of data centers mostly by renewable sources, and to understand what portfolio of renewable energy is most effective.

Proceedings ArticleDOI
Qian Zhu1, Teresa Tung1
24 Jun 2012
TL;DR: Using the proposed interference model to optimize the cloud provider's metric (here the number of successfully executed applications) to realize better workload placement decisions and thereby maintaining the user's application QoS.
Abstract: Cloud computing offers users the ability to access large pools of computational and storage resources on-demand without the burden of managing and maintaining their own IT assets. Today's cloud providers charge users based upon the amount of resources used or reserved, with only minimal guarantees of the quality-of-service (QoS) experienced byte users applications. As virtualization technologies proliferate among cloud providers, consolidating multiple user applications onto multi-core servers increases revenue and improves resource utilization. However, consolidation introduces performance interference between co-located workloads, which significantly impacts application QoS. A critical requirement for effective consolidation is to be able to predict the impact of application performance in the presence of interference from on-chip resources, e.g., CPU and last-level cache (LLC)/memory bandwidth sharing, to storage devices and network bandwidth contention. In this work, we propose an interference model which predicts the application QoS metric. The key distinctive feature is the consideration of time-variant inter-dependency among different levels of resource interference. We use applications from a test suite and SPECWeb2005 to illustrate the effectiveness of our model and an average prediction error of less than 8% is achieved. Furthermore, we demonstrate using the proposed interference model to optimize the cloud provider's metric (here the number of successfully executed applications) to realize better workload placement decisions and thereby maintaining the user's application QoS.

Proceedings ArticleDOI
16 Apr 2012
TL;DR: This paper uses data traces obtained from a real data center to develop capabilities of characterizing and predicting workload on the Virtual Machines, and introduces a method based on Hidden Markov Modeling to characterize the temporal correlations in the discovered VM clusters and to predict variations of workload patterns.
Abstract: Cloud computing promises high scalability, flexibility and cost-effectiveness to satisfy emerging computing requirements. To efficiently provision computing resources in the cloud, system administrators need the capabilities of characterizing and predicting workload on the Virtual Machines (VMs). In this paper, we use data traces obtained from a real data center to develop such capabilities. First, we search for repeatable workload patterns by exploring cross-VM workload correlations resulted from the dependencies among applications running on different VMs. Treating workload data samples as time series, we develop a co-clustering technique to identify groups of VMs that frequently exhibit correlated workload patterns, and also the time periods in which these VM groups are active. Then, we introduce a method based on Hidden Markov Modeling (HMM) to characterize the temporal correlations in the discovered VM clusters and to predict variations of workload patterns. The experimental results show that our method can not only help better understand group-level workload characteristics, but also make more accurate predictions on workload changes in a cloud.

Journal ArticleDOI
TL;DR: A dynamic capacity management policy, AutoScale, is introduced that greatly reduces the number of servers needed in data centers driven by unpredictable, time-varying load, while meeting response time SLAs and robustness.
Abstract: Energy costs for data centers continue to rise, already exceeding $15 billion yearly. Sadly much of this power is wasted. Servers are only busy 10--30p of the time on average, but they are often left on, while idle, utilizing 60p or more of peak power when in the idle state.We introduce a dynamic capacity management policy, AutoScale, that greatly reduces the number of servers needed in data centers driven by unpredictable, time-varying load, while meeting response time SLAs. AutoScale scales the data center capacity, adding or removing servers as needed. AutoScale has two key features: (i) it autonomically maintains just the right amount of spare capacity to handle bursts in the request rate; and (ii) it is robust not just to changes in the request rate of real-world traces, but also request size and server efficiency.We evaluate our dynamic capacity management approach via implementation on a 38-server multi-tier data center, serving a web site of the type seen in Facebook or Amazon, with a key-value store workload. We demonstrate that AutoScale vastly improves upon existing dynamic capacity management policies with respect to meeting SLAs and robustness.

Patent
28 Mar 2012
TL;DR: In this paper, a system level scheme for networking of implantable devices, electronic patch devices/sensors coupled to the body, and wearable sensors/devices with cellular telephone/mobile devices, peripheral devices and remote servers is described.
Abstract: A system level scheme for networking of implantable devices, electronic patch devices/sensors coupled to the body, and wearable sensors/devices with cellular telephone/mobile devices, peripheral devices and remote servers is described.

Proceedings ArticleDOI
16 Apr 2012
TL;DR: This paper focuses on horizontal elasticity, the ability of the infrastructure to add or remove virtual machines allocated to a service deployed in the cloud, and builds two adaptive proactive controllers that estimate the future load on a service using queuing theory.
Abstract: Cloud elasticity is the ability of the cloud infrastructure to rapidly change the amount of resources allocated to a service in order to meet the actual varying demands on the service while enforcing SLAs. In this paper, we focus on horizontal elasticity, the ability of the infrastructure to add or remove virtual machines allocated to a service deployed in the cloud. We model a cloud service using queuing theory. Using that model we build two adaptive proactive controllers that estimate the future load on a service. We explore the different possible scenarios for deploying a proactive elasticity controller coupled with a reactive elasticity controller in the cloud. Using simulation with workload traces from the FIFA world-cup web servers, we show that a hybrid controller that incorporates a reactive controller for scale up coupled with our proactive controllers for scale down decisions reduces SLA violations by a factor of 2 to 10 compared to a regression based controller or a completely reactive controller.

Proceedings ArticleDOI
01 Dec 2012
TL;DR: How Internet of Things and Cloud computing can work together can address the Big Data issues is described and a prototype model for providing sensing as a service on cloud is proposed.
Abstract: Internet of Things (IoT) is a concept that envisions all objects around us as part of internet. IoT coverage is very wide and include variety of objects like smart phones, tablets, digital cameras, sensors, etc. Once all these devices are connected with each other, they enable more and more smart processes and services that support our basic needs, economies, environment and health. Such enormous number of devices connected to internet provides many kinds of services and produce huge amount of data and information. Cloud computing is a model for on-demand access to a shared pool of configurable resources (e.g. compute, networks, servers, storage, applications, services, and software) that can be easily provisioned as Infrastructure (IaaS), software and applications (SaaS). Cloud based platforms help to connect to the things (IaaS) around us so that we can access anything at any time and any place in a user friendly manner using customized portals and in built applications (SaaS). Hence, cloud acts as a front end to access Internet of Things. Applications that interact with devices like sensors have special requirements of massive storage to storage big data, huge computation power to enable the real time processing of the data, and high speed network to stream audio or video. In this paper, we describe how Internet of Things and Cloud computing can work together can address the Big Data issues. We also illustrate about Sensing as a service on cloud using few applications like Augmented Reality, Agriculture and Environment monitoring. Finally, we also propose a prototype model for providing sensing as a service on cloud.

Proceedings ArticleDOI
03 Dec 2012
TL;DR: This paper investigates the different algorithms proposed to resolve the issue of load balancing and task scheduling in Cloud Computing and discusses and compares these algorithms to provide an overview of the latest approaches in the field.
Abstract: Load Balancing is essential for efficient operations indistributed environments. As Cloud Computing is growingrapidly and clients are demanding more services and betterresults, load balancing for the Cloud has become a veryinteresting and important research area. Many algorithms weresuggested to provide efficient mechanisms and algorithms forassigning the client's requests to available Cloud nodes. Theseapproaches aim to enhance the overall performance of the Cloudand provide the user more satisfying and efficient services. Inthis paper, we investigate the different algorithms proposed toresolve the issue of load balancing and task scheduling in CloudComputing. We discuss and compare these algorithms to providean overview of the latest approaches in the field.

Proceedings ArticleDOI
25 Mar 2012
TL;DR: The design of YouTube video delivery system consists of a “flat” video id space, multiple DNS namespaces reflecting a multi-layered logical organization of video servers, and a 3-tier physical cache hierarchy.
Abstract: We deduce key design features behind the YouTube video delivery system by building a distributed active measurement infrastructure, and collecting and analyzing a large volume of video playback logs, DNS mappings and latency data. We find that the design of YouTube video delivery system consists of three major components: a “flat” video id space, multiple DNS namespaces reflecting a multi-layered logical organization of video servers, and a 3-tier physical cache hierarchy. We also uncover that YouTube employs a set of sophisticated mechanisms to handle video delivery dynamics such as cache misses and load sharing among its distributed cache locations and data centers.

Proceedings ArticleDOI
03 Mar 2012
TL;DR: Tarazu, a suite of optimizations to improve MapReduce performance on heterogeneous clusters, is presented, showing that Tarazu significantly improves performance over a baseline of Hadoop with straightforward tuning for hardware heterogeneity.
Abstract: Data center-scale clusters are evolving towards heterogeneous hardware for power, cost, differentiated price-performance, and other reasons. MapReduce is a well-known programming model to process large amount of data on data center-scale clusters. Most MapReduce implementations have been designed and optimized for homogeneous clusters. Unfortunately, these implementations perform poorly on heterogeneous clusters (e.g., on a 90-node cluster that contains 10 Xeon-based servers and 80 Atom-based servers, Hadoop performs worse than on 10-node Xeon-only or 80-node Atom-only homogeneous sub-clusters for many of our benchmarks). This poor performance remains despite previously proposed optimizations related to management of straggler tasks. In this paper, we address MapReduce's poor performance on heterogeneous clusters. Our first contribution is that the poor performance is due to two key factors: (1) the non-intuitive effect that MapReduce's built-in load balancing results in excessive and bursty network communication during the Map phase, and (2) the intuitive effect that the heterogeneity amplifies load imbalance in the Reduce computation. Our second contribution is Tarazu, a suite of optimizations to improve MapReduce performance on heterogeneous clusters. Tarazu consists of (1) Communication-Aware Load Balancing of Map computation (CALB) across the nodes, (2) Communication-Aware Scheduling of Map computation (CAS) to avoid bursty network traffic and (3) Predictive Load Balancing of Reduce computation (PLB) across the nodes. Using the above 90-node cluster, we show that Tarazu significantly improves performance over a baseline of Hadoop with straightforward tuning for hardware heterogeneity.

Journal ArticleDOI
TL;DR: The proposed model is not to introduce a new schema in contrast to IFC but to harness the capability of IFC XML and or possibly engage with using Simplified Markup Language (SML) subsets of eXtensible Mark up Language (XML) for exchanging partial data to design an integrated platform that would enhance the BIM usability experience for various disciplines in making key design decisions at a relatively early design stage.