Institution
Hewlett-Packard
Company•Palo Alto, California, United States•
About: Hewlett-Packard is a company organization based out in Palo Alto, California, United States. It is known for research contribution in the topics: Signal & Substrate (printing). The organization has 34663 authors who have published 59808 publications receiving 1467218 citations. The organization is also known as: Hewlett Packard & Hewlett-Packard Company.
Papers published on a yearly basis
Papers
More filters
•
01 Jan 2002TL;DR: Zipf's law governs many features of the Internet, and web caching strategies are formulated to account for a Zipf distribution in the number of requests for webpages.
Abstract: Zipf's law governs many features of the Internet. Observations of Zipf distributions, while interesting in and of themselves, have strong implications for the design and function of the Internet. The connectivity of Internet routers influences the robustness of the network while the distribution in the number of email contacts affects the spread of email viruses. Even web caching strategies are formulated to account for a Zipf distribution in the number of requests for webpages.
755 citations
••
01 May 1995TL;DR: Compared with previous quadtree-based algorithms for quality mesh generation, the Delaunay refinement approach is much simpler and generally produces meshes with fewer triangles.
Abstract: We present a simple new algorithm for triangulating polygons and planar straightline graphs, It provides "shape" and "size" guarantees: ?All triangles have a bounded aspect ratio.?The number of triangles is within a constant factor of optimal. Such "quality" triangulations are desirable as meshes for the finite element method, in which the running time generally increases with the number of triangles, and where the convergence and stability may be hurt by very skinny triangles. The technique we use-successive refinement of a Delaunay triangulation-extends a mesh generation technique of Chew by allowing triangles of varying sizes. Compared with previous quadtree-based algorithms for quality mesh generation, the Delaunay refinement approach is much simpler and generally produces meshes with fewer triangles. We also discuss an implementation of the algorithm and evaluate its performance on a variety of inputs.
752 citations
••
TL;DR: It becomes critically important to study how the current approaches to standardization in this area can be improved, and better understand the opportunities for the research community to contribute to the IoT field.
Abstract: Technologies to support the Internet of Things are becoming more important as the need to better understand our environments and make them smart increases. As a result it is predicted that intelligent devices and networks, such as WSNs, will not be isolated, but connected and integrated, composing computer networks. So far, the IP-based Internet is the largest network in the world; therefore, there are great strides to connect WSNs with the Internet. To this end, the IETF has developed a suite of protocols and open standards for accessing applications and services for wireless resource constrained networks. However, many open challenges remain, mostly due to the complex deployment characteristics of such systems and the stringent requirements imposed by various services wishing to make use of such complex systems. Thus, it becomes critically important to study how the current approaches to standardization in this area can be improved, and at the same time better understand the opportunities for the research community to contribute to the IoT field. To this end, this article presents an overview of current standards and research activities in both industry and academia.
744 citations
••
TL;DR: It is found that improvements in the caching architecture of the World Wide Web are changing the workloads of Web servers, but major improvements to that architecture are still necessary.
Abstract: This article presents a detailed workload characterization study of the 1998 World Cup Web site. Measurements from this site were collected over a three-month period. During this time the site received 1.35 billion requests, making this the largest Web workload analyzed to date. By examining this extremely busy site and through comparison with existing characterization studies, we are able to determine how Web server workloads are evolving. We find that improvements in the caching architecture of the World Wide Web are changing the workloads of Web servers, but major improvements to that architecture are still necessary. In particular, we uncover evidence that a better consistency mechanism is required for World Wide Web caches.
743 citations
•
10 Apr 2005TL;DR: This paper examines a theoretic thermodynamic formulation that uses information about steady state hot spots and cold spots in the data center and develops real-world scheduling algorithms, and develops an alternate approach to address the problem of heat management through temperature-aware workload placement.
Abstract: Trends towards consolidation and higher-density computing configurations make the problem of heat management one of the critical challenges in emerging data centers Conventional approaches to addressing this problem have focused at the facilities level to develop new cooling technologies or optimize the delivery of cooling In contrast to these approaches, our paper explores an alternate dimension to address this problem, namely a systems-level solution to control the heat generation through temperature-aware workload placement
We first examine a theoretic thermodynamic formulation that uses information about steady state hot spots and cold spots in the data center and develop real-world scheduling algorithms Based on the insights from these results, we develop an alternate approach Our new approach leverages the non-intuitive observation that the source of cooling inefficiencies can often be in locations spatially uncorrelated with its manifested consequences; this enables additional energy savings Overall, our results demonstrate up to a factor of two reduction in annual data center cooling costs over location-agnostic workload distribution, purely through software optimizations without the need for any costly capital investment
740 citations
Authors
Showing all 34676 results
Name | H-index | Papers | Citations |
---|---|---|---|
Andrew White | 149 | 1494 | 113874 |
Stephen R. Forrest | 148 | 1041 | 111816 |
Rafi Ahmed | 146 | 633 | 93190 |
Leonidas J. Guibas | 124 | 691 | 79200 |
Chenming Hu | 119 | 1296 | 57264 |
Robert E. Tarjan | 114 | 400 | 67305 |
Hong-Jiang Zhang | 112 | 461 | 49068 |
Ching-Ping Wong | 106 | 1128 | 42835 |
Guillermo Sapiro | 104 | 667 | 70128 |
James R. Heath | 103 | 425 | 58548 |
Arun Majumdar | 102 | 459 | 52464 |
Luca Benini | 101 | 1453 | 47862 |
R. Stanley Williams | 100 | 605 | 46448 |
David M. Blei | 98 | 378 | 111547 |
Wei-Ying Ma | 97 | 464 | 40914 |