NUS-PT: exploiting parallel texts for word sense disambiguation in the English all-words tasks
SemEval '07 Proceedings of the 4th International Workshop on Semantic Evaluations
UBC-ALM: combining k-NN with SVD for WSD
SemEval '07 Proceedings of the 4th International Workshop on Semantic Evaluations
Overview of the CLEF 2005 multilingual question answering track
CLEF'05 Proceedings of the 6th international conference on Cross-Language Evalution Forum: accessing Multilingual Information Repositories
Question answering pilot task at CLEF 2004
CLEF'04 Proceedings of the 5th conference on Cross-Language Evaluation Forum: multilingual Information Access for Text, Speech and Images
Overview of the CLEF 2006 multilingual question answering track
CLEF'06 Proceedings of the 7th international conference on Cross-Language Evaluation Forum: evaluation of multilingual and multi-modal information retrieval
An application of automated reasoning in natural language question answering
AI Communications - Practical Aspects of Automated Reasoning
Overview of the answer validation exercise 2008
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
Using AliQAn in monolingual QA@CLEF 2008
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
IdSay: question answering for Portuguese
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
Using answer retrieval patterns to answer Portuguese questions
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
UAIC participation at QA@CLEF2008
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
Integrating logic forms and anaphora resolution in the AliQAn system
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
Analyzing the use of non-overlap features for supervised answer validation
CLEF'08 Proceedings of the 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access
Question answering for Portuguese: how much is needed?
SBIA'10 Proceedings of the 20th Brazilian conference on Advances in artificial intelligence
Learning to select the correct answer in multi-stream question answering
Information Processing and Management: an International Journal
PROPOR'10 Proceedings of the 9th international conference on Computational Processing of the Portuguese Language
Latent association analysis of document pairs
Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining
Question answering at the cross-language evaluation forum 2003---2010
Language Resources and Evaluation
Hi-index | 0.00 |
The QA campaign at CLEF 2008 [1], was mainly the same as that proposed last year. The results and the analyses reported by last year's participants suggested that the changes introduced in the previous campaign had led to a drop in systems' performance. So for this year's competition it has been decided to practically replicate last year's exercise. Following last year's experience some QA pairs were grouped in clusters. Every cluster was characterized by a topic (not given to participants). The questions from a cluster contained coreferences between one of them and the others. Moreover, as last year, the systems were given the possibility to search for answers in Wikipedia as document corpus beside the usual newswire collection. In addition to the main task, three additional exercises were offered, namely the Answer Validation Exercise (AVE), the Question Answering on Speech Transcriptions (QAST), which continued last year's successful pilots, together with the new Word Sense Disambiguation for Question Answering (QA-WSD). As general remark, it must be said that the main task still proved to be very challenging for participating systems. As a kind of shallow comparison with last year's results the best overall accuracy dropped significantly from 42% to 19% in the multi-lingual subtasks, but increased a little in the monolingual sub-tasks, going from 54% to 63%.