scispace - formally typeset
Open AccessProceedings ArticleDOI

Efficient peer-to-peer keyword searching

TLDR
A distributed search engine based on a distributed hash table is designed and analyzed and the simulation results predict that the search engine can answer an average query in under one second, using under one kilobyte of bandwidth.
Abstract
The recent file storage applications built on top of peer-to-peer distributed hash tables lack search capabilities. We believe that search is an important part of any document publication system. To that end, we have designed and analyzed a distributed search engine based on a distributed hash table. Our simulation results predict that our search engine can answer an average query in under one second, using under one kilobyte of bandwidth.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Tag Indexed DHT for Scalable Search Infrastructure in Social NetworkApplications

TL;DR: T-DHT behaves as a structured DHT when publishing "tag, social-object" associations, and as an unstructured filter driven network when searching for the social-objects by means of any tag combination.
Proceedings ArticleDOI

Scalable Semantic Search with Hybrid Concept Index over Structure Peer-to-Peer Network

TL;DR: Simulation experiments show that keyword search with the approach proposed in this paper is much less on bandwidth costs and much higher on retrieval perform than that based on standard inverted index by keywords.
Journal ArticleDOI

Tree-Based Index Overlay in Hybrid Peer-to-Peer Systems

TL;DR: The mathematical analysis shows that the keyword search based on semi-structured P2P overlay can improve the search performance, reducing the message traffic and maintenance costs.
Dissertation

Trust-aware information retrieval in peer-to-peer environments

Ye Zhang
TL;DR: The work presented in this dissertation provides the first integrated framework for trust-aware Information Retrieval in P2P environments, which can retrieve not only relevant but also trustworthy documents and reduce the possibility of untrustworthy documents in the top-ranked result list.
Book ChapterDOI

LINP: supporting similarity search in unstructured peer-to-peer networks

TL;DR: This paper designs an efficient indexmechanism, named Linking Identical Neighborly Partitions (LINP), which takes advantage of both space partitioning and routing indices techniques and evaluates the proposed scheme over various data sets, and experimental results show the efficacy of the approach.
References
More filters
Journal ArticleDOI

The anatomy of a large-scale hypertextual Web search engine

TL;DR: This paper provides an in-depth description of Google, a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and looks at the problem of how to effectively deal with uncontrolled hypertext collections where anyone can publish anything they want.
Proceedings Article

The PageRank Citation Ranking : Bringing Order to the Web

TL;DR: This paper describes PageRank, a mathod for rating Web pages objectively and mechanically, effectively measuring the human interest and attention devoted to them, and shows how to efficiently compute PageRank for large numbers of pages.
Journal Article

The Anatomy of a Large-Scale Hypertextual Web Search Engine.

Sergey Brin, +1 more
- 01 Jan 1998 - 
TL;DR: Google as discussed by the authors is a prototype of a large-scale search engine which makes heavy use of the structure present in hypertext and is designed to crawl and index the Web efficiently and produce much more satisfying search results than existing systems.
Proceedings ArticleDOI

Chord: A scalable peer-to-peer lookup service for internet applications

TL;DR: Results from theoretical analysis, simulations, and experiments show that Chord is scalable, with communication cost and the state maintained by each node scaling logarithmically with the number of Chord nodes.
Journal ArticleDOI

Space/time trade-offs in hash coding with allowable errors

TL;DR: Analysis of the paradigm problem demonstrates that allowing a small number of test messages to be falsely identified as members of the given set will permit a much smaller hash area to be used without increasing reject time.
Related Papers (5)