Home News About Us Contact Contributors Disclaimer Privacy Policy Help FAQ

Home
Search
Quick Search
Advanced
Fulltext
Browse
Collections
Persons
My eDoc
Session History
Login
Name:
Password:
Documentation
Help
Support Wiki
Direct access to
document ID:


          Display Documents



  history
ID: 314435.0, MPI für Informatik / Databases and Information Systems Group
A Reproducible Benchmark for P2P Retrieval
Authors:Neumann, Thomas; Bender, Matthias; Michel, Sebastian; Weikum, Gerhard
Editors:Bonnet, Philippe; Manolescu, Ioana
Language:English
Publisher:ACM
Place of Publication:New York, USA
Date of Publication (YYYY-MM-DD):2006
Title of Proceedings:Proceedings of the 1st International Workshop on Performance and Evaluation of Data Management Systems, ExpDB 2006, in cooperation with ACM SIGMOD
Start Page:1
End Page:8
Place of Conference/Meeting:Chicago, Illinois, USA
(Start) Date of Conference/Meeting
 (YYYY-MM-DD):
2006-06-30
Review Status:not specified
Audience:Experts Only
Intended Educational Use:No
Abstract / Description:With the growing popularity of information retrieval (IR)
in distributed systems and in particular {P2P} Web search, a
huge number of protocols and prototypes have been introduced
in the literature. However, nearly every paper considers
a different benchmark for its experimental evaluation,
rendering their mutual comparison and the quantification of
performance improvements an impossible task.
We present a standardized, general purpose benchmark
for {P2P IR} systems that finally makes this possible. We
start by presenting a detailed requirement analysis for such
a standardized benchmark framework that allows for reproducible
and comparable experimental setups without sacrificing
flexibility to suit different system models. We further
suggest Wikipedia as a publicly-available and all-purpose
document corpus and finally introduce a simple but yet flexible
clustering strategy that assigns the Wikipedia articles as
documents to an arbitrary number of peers. After proposing
a standardized, real-world query set as the benchmark
workload, we review the metrics to evaluate the benchmark
results and present an example benchmark run for our fullyimplemented
{P2P} Web search prototype {MINERVA}.
Last Change of the Resource (YYYY-MM-DD):2007-04-27
External Publication Status:published
Document Type:Conference-Paper
Communicated by:Gerhard Weikum
Affiliations:MPI für Informatik/Databases and Information Systems Group
Identifiers:ISBN:1-59593-463-4
LOCALID:C1256DBF005F876D-B1D251BC8260E5E9C12571B80053D231-...
Full Text:
You have privileges to view the following file(s):
NeumannBMW06.pdf  [159,00 Kb] [Comment:file from upload service]  
 
The scope and number of records on eDoc is subject to the collection policies defined by each institute - see "info" button in the collection browse view.