Design and Evaluation of an IR-Benchmark for SPARQL Queries with Fulltext Conditions

From WikiPapers
Jump to: navigation, search

Design and Evaluation of an IR-Benchmark for SPARQL Queries with Fulltext Conditions is a 2012 conference paper written in English by Mishra A., Gurajada S., Theobald M. and published in International Conference on Information and Knowledge Management, Proceedings.

[edit] Abstract

In this paper, we describe our goals in introducing a new, annotated benchmark collection, with which we aim to bridge the gap between the fundamentally different aspects that are involved in querying both structured and unstructured data. This semantically rich collection, captured in a unified XML format, combines components (unstructured text, semistructured infoboxes, and category structure) from 3.1 Million Wikipedia articles with highly structured RDF properties from both DBpedia and YAGO2. The new collection serves as the basis of the INEX 2012 Ad-hoc, Faceted Search, and Jeopardy retrieval tasks. With a focus on the new Jeopardy task, we particularly motivate the usage of the collection for question-answering (QA) style retrieval settings, which we also exemplify by introducing a set of 90 QA-style benchmark queries which come shipped in a SPARQL-based query format that has been extended by fulltext filter conditions.

[edit] References

This section requires expansion. Please, help!

Cited by

Probably, this publication is cited by others, but there are no articles available for them in WikiPapers.