For the current REF see the REF 2021 website REF 2021 logo

Output details

11 - Computer Science and Informatics

University of Edinburgh

Return to search Previous output Next output
Output 0 of 0 in the submission
Output title

Repeatable and reliable search system evaluation using crowdsourcing

Type
E - Conference contribution
Name of conference/published proceedings
Proceedings of the 34th international ACM SIGIR conference on Research and development in Information Retrieval
Volume number
-
Issue number
-
First page of article
923
ISSN of proceedings
-
Year of publication
2011
Number of additional authors
6
Additional information

<16> Originality: First-ever demonstration of the repeatability and reliability of crowdsourced evaluation of semantic-web search-engines using multiple trials and comparing with expert judgements.

Significance: Describes an ongoing evaluation campaign of semweb search system performance with submissions from academic and commercial participants. Repeatability enables comparison of performance across time for deployed search engines. Supports work towards fully-automated develop-test-evaluate loop by showing high agreement between distinct judge pools separated in time. Subsequent competitions have reused this methodology (http://www.websemanticsjournal.org/index.php/ps/article/view/336).

Rigour: Demonstrates reliability using multiple evaluation and relevance metrics. Introduces objective criteria for rejecting unreliable crowdsourced judges. Acceptance rate for this SIGIR was 19.9%.

Interdisciplinary
-
Cross-referral requested
-
Research group
D - Institute for Language, Cognition & Computation
Citation count
10
Proposed double-weighted
No
Double-weighted statement
-
Reserve for a double-weighted output
No
Non-English
No
English abstract
-