TY - GEN
T1 - Overview of ResPubliQA 2009
T2 - 10th Workshop of the Cross-Language Evaluation Forum, CLEF 2009
AU - Peñas, Anselmo
AU - Forner, Pamela
AU - Sutcliffe, Richard
AU - Rodrigo, Álvaro
AU - Forǎscu, Corina
AU - Alegria, Iñaki
AU - Giampiccolo, Danilo
AU - Moreau, Nicolas
AU - Osenova, Petya
PY - 2010
Y1 - 2010
N2 - This paper describes the first round of ResPubliQA, a Question Answering (QA) evaluation task over European legislation, proposed at the Cross Language Evaluation Forum (CLEF) 2009. The exercise consists of extracting a relevant paragraph of text that satisfies completely the information need expressed by a natural language question. The general goals of this exercise are (i) to study if the current QA technologies tuned for newswire collections and Wikipedia can be adapted to a new domain (law in this case); (ii) to move to a more realistic scenario, considering people close to law as users, and paragraphs as system output; (iii) to compare current QA technologies with pure Information Retrieval (IR) approaches; and (iv) to introduce in QA systems the Answer Validation technologies developed in the past three years. The paper describes the task in more detail, presenting the different types of questions, the methodology for the creation of the test sets and the new evaluation measure, and analyzing the results obtained by systems and the more successful approaches. Eleven groups participated with 28 runs. In addition, we evaluated 16 baseline runs (2 per language) based only in pure IR approach, for comparison purposes. Considering accuracy, scores were generally higher than in previous QA campaigns.
AB - This paper describes the first round of ResPubliQA, a Question Answering (QA) evaluation task over European legislation, proposed at the Cross Language Evaluation Forum (CLEF) 2009. The exercise consists of extracting a relevant paragraph of text that satisfies completely the information need expressed by a natural language question. The general goals of this exercise are (i) to study if the current QA technologies tuned for newswire collections and Wikipedia can be adapted to a new domain (law in this case); (ii) to move to a more realistic scenario, considering people close to law as users, and paragraphs as system output; (iii) to compare current QA technologies with pure Information Retrieval (IR) approaches; and (iv) to introduce in QA systems the Answer Validation technologies developed in the past three years. The paper describes the task in more detail, presenting the different types of questions, the methodology for the creation of the test sets and the new evaluation measure, and analyzing the results obtained by systems and the more successful approaches. Eleven groups participated with 28 runs. In addition, we evaluated 16 baseline runs (2 per language) based only in pure IR approach, for comparison purposes. Considering accuracy, scores were generally higher than in previous QA campaigns.
UR - http://www.scopus.com/inward/record.url?scp=78049344440&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-15754-7_21
DO - 10.1007/978-3-642-15754-7_21
M3 - Conference contribution
AN - SCOPUS:78049344440
SN - 364215753X
SN - 9783642157530
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 174
EP - 196
BT - Multilingual Information Access Evaluation I
Y2 - 30 September 2009 through 2 October 2009
ER -