scispace - formally typeset
Open AccessProceedings ArticleDOI

Efficient peer-to-peer keyword searching

TLDR
A distributed search engine based on a distributed hash table is designed and analyzed and the simulation results predict that the search engine can answer an average query in under one second, using under one kilobyte of bandwidth.
Abstract
The recent file storage applications built on top of peer-to-peer distributed hash tables lack search capabilities. We believe that search is an important part of any document publication system. To that end, we have designed and analyzed a distributed search engine based on a distributed hash table. Our simulation results predict that our search engine can answer an average query in under one second, using under one kilobyte of bandwidth.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Probabilistic Data Structures in Adversarial Environments

TL;DR: This work provides a provable security treatment of probabilistic data structures in adversarial environments with a syntax that captures a wide variety of in-use structures, and security notions support development of error bounds in the presence of powerful attacks.
Proceedings ArticleDOI

The dynamic cuckoo filter

TL;DR: The dynamic cuckoo filter (DCF) is proposed to support reliable delete operation and elastic capacity for dynamic set representation and membership testing and comprehensive experiment results demonstrate the efficiency of the DCF design compared to existing schemes.
Proceedings ArticleDOI

Optimized Inverted List Assignment in Distributed Search Engine Architectures

TL;DR: This work analyzes search engine query traces in order to optimize the assignment of index data to the nodes in the system, such that terms frequently occurring together in queries are also often collocated on the same node.
Proceedings ArticleDOI

Keyword fusion to support efficient keyword-based search in peer-to-peer file sharing

TL;DR: This paper proposes a set of mechanisms to provide a scalable keyword-based file search in distributed hash table (DHT)-based P2P systems called keyword fusion, which adaptively unburdens the peers overloaded with excessive storage consumptions due to common keywords and reduces network bandwidth consumption.
Proceedings ArticleDOI

Bringing efficient advanced queries to distributed hash tables

TL;DR: This paper introduces new efficient, scalable, and completely distributed methods that strive to keep resource consumption by queries and index information as low as possible and describes how to improve the handling of multiple subqueries combined through Boolean set operators.
References
More filters
Journal ArticleDOI

The anatomy of a large-scale hypertextual Web search engine

TL;DR: This paper provides an in-depth description of Google, a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and looks at the problem of how to effectively deal with uncontrolled hypertext collections where anyone can publish anything they want.
Proceedings Article

The PageRank Citation Ranking : Bringing Order to the Web

TL;DR: This paper describes PageRank, a mathod for rating Web pages objectively and mechanically, effectively measuring the human interest and attention devoted to them, and shows how to efficiently compute PageRank for large numbers of pages.
Journal Article

The Anatomy of a Large-Scale Hypertextual Web Search Engine.

Sergey Brin, +1 more
- 01 Jan 1998 - 
TL;DR: Google as discussed by the authors is a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and is designed to crawl and index the Web efficiently and produce much more satisfying search results than existing systems.
Proceedings ArticleDOI

Chord: A scalable peer-to-peer lookup service for internet applications

TL;DR: Results from theoretical analysis, simulations, and experiments show that Chord is scalable, with communication cost and the state maintained by each node scaling logarithmically with the number of Chord nodes.
Journal ArticleDOI

Space/time trade-offs in hash coding with allowable errors

TL;DR: Analysis of the paradigm problem demonstrates that allowing a small number of test messages to be falsely identified as members of the given set will permit a much smaller hash area to be used without increasing reject time.
Related Papers (5)