scispace - formally typeset
Search or ask a question
Institution

Hewlett-Packard

CompanyPalo Alto, California, United States
About: Hewlett-Packard is a company organization based out in Palo Alto, California, United States. It is known for research contribution in the topics: Signal & Substrate (printing). The organization has 34663 authors who have published 59808 publications receiving 1467218 citations. The organization is also known as: Hewlett Packard & Hewlett-Packard Company.


Papers
More filters
Proceedings Article
01 Jan 2002
TL;DR: Zipf's law governs many features of the Internet, and web caching strategies are formulated to account for a Zipf distribution in the number of requests for webpages.
Abstract: Zipf's law governs many features of the Internet. Observations of Zipf distributions, while interesting in and of themselves, have strong implications for the design and function of the Internet. The connectivity of Internet routers influences the robustness of the network while the distribution in the number of email contacts affects the spread of email viruses. Even web caching strategies are formulated to account for a Zipf distribution in the number of requests for webpages.

755 citations

Journal ArticleDOI
Jim Ruppert1
01 May 1995
TL;DR: Compared with previous quadtree-based algorithms for quality mesh generation, the Delaunay refinement approach is much simpler and generally produces meshes with fewer triangles.
Abstract: We present a simple new algorithm for triangulating polygons and planar straightline graphs, It provides "shape" and "size" guarantees: ?All triangles have a bounded aspect ratio.?The number of triangles is within a constant factor of optimal. Such "quality" triangulations are desirable as meshes for the finite element method, in which the running time generally increases with the number of triangles, and where the convergence and stability may be hurt by very skinny triangles. The technique we use-successive refinement of a Delaunay triangulation-extends a mesh generation technique of Chew by allowing triangles of varying sizes. Compared with previous quadtree-based algorithms for quality mesh generation, the Delaunay refinement approach is much simpler and generally produces meshes with fewer triangles. We also discuss an implementation of the algorithm and evaluate its performance on a variety of inputs.

752 citations

Journal ArticleDOI
TL;DR: It becomes critically important to study how the current approaches to standardization in this area can be improved, and better understand the opportunities for the research community to contribute to the IoT field.
Abstract: Technologies to support the Internet of Things are becoming more important as the need to better understand our environments and make them smart increases. As a result it is predicted that intelligent devices and networks, such as WSNs, will not be isolated, but connected and integrated, composing computer networks. So far, the IP-based Internet is the largest network in the world; therefore, there are great strides to connect WSNs with the Internet. To this end, the IETF has developed a suite of protocols and open standards for accessing applications and services for wireless resource constrained networks. However, many open challenges remain, mostly due to the complex deployment characteristics of such systems and the stringent requirements imposed by various services wishing to make use of such complex systems. Thus, it becomes critically important to study how the current approaches to standardization in this area can be improved, and at the same time better understand the opportunities for the research community to contribute to the IoT field. To this end, this article presents an overview of current standards and research activities in both industry and academia.

744 citations

Journal ArticleDOI
TL;DR: It is found that improvements in the caching architecture of the World Wide Web are changing the workloads of Web servers, but major improvements to that architecture are still necessary.
Abstract: This article presents a detailed workload characterization study of the 1998 World Cup Web site. Measurements from this site were collected over a three-month period. During this time the site received 1.35 billion requests, making this the largest Web workload analyzed to date. By examining this extremely busy site and through comparison with existing characterization studies, we are able to determine how Web server workloads are evolving. We find that improvements in the caching architecture of the World Wide Web are changing the workloads of Web servers, but major improvements to that architecture are still necessary. In particular, we uncover evidence that a better consistency mechanism is required for World Wide Web caches.

743 citations

Proceedings Article
10 Apr 2005
TL;DR: This paper examines a theoretic thermodynamic formulation that uses information about steady state hot spots and cold spots in the data center and develops real-world scheduling algorithms, and develops an alternate approach to address the problem of heat management through temperature-aware workload placement.
Abstract: Trends towards consolidation and higher-density computing configurations make the problem of heat management one of the critical challenges in emerging data centers Conventional approaches to addressing this problem have focused at the facilities level to develop new cooling technologies or optimize the delivery of cooling In contrast to these approaches, our paper explores an alternate dimension to address this problem, namely a systems-level solution to control the heat generation through temperature-aware workload placement We first examine a theoretic thermodynamic formulation that uses information about steady state hot spots and cold spots in the data center and develop real-world scheduling algorithms Based on the insights from these results, we develop an alternate approach Our new approach leverages the non-intuitive observation that the source of cooling inefficiencies can often be in locations spatially uncorrelated with its manifested consequences; this enables additional energy savings Overall, our results demonstrate up to a factor of two reduction in annual data center cooling costs over location-agnostic workload distribution, purely through software optimizations without the need for any costly capital investment

740 citations


Authors

Showing all 34676 results

NameH-indexPapersCitations
Andrew White1491494113874
Stephen R. Forrest1481041111816
Rafi Ahmed14663393190
Leonidas J. Guibas12469179200
Chenming Hu119129657264
Robert E. Tarjan11440067305
Hong-Jiang Zhang11246149068
Ching-Ping Wong106112842835
Guillermo Sapiro10466770128
James R. Heath10342558548
Arun Majumdar10245952464
Luca Benini101145347862
R. Stanley Williams10060546448
David M. Blei98378111547
Wei-Ying Ma9746440914
Network Information
Related Institutions (5)
IBM
253.9K papers, 7.4M citations

94% related

Samsung
163.6K papers, 2M citations

90% related

Carnegie Mellon University
104.3K papers, 5.9M citations

90% related

Microsoft
86.9K papers, 4.1M citations

90% related

Bell Labs
59.8K papers, 3.1M citations

89% related

Performance
Metrics
No. of papers from the Institution in previous years
YearPapers
20231
202223
2021240
20201,028
20191,269
2018964