Building a reusable test collection for question answering

被引:20
作者
Lin, J [1 ]
Katz, B
机构
[1] Univ Maryland, Coll Informat Studies, Inst Adv Comp Studies, College Pk, MD 20742 USA
[2] MIT, Comp Sci & Artificial Intelligence Lab, Cambridge, MA 02139 USA
来源
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY | 2006年 / 57卷 / 07期
关键词
D O I
10.1002/asi.20348
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In contrast to traditional information retrieval systems, which return ranked lists of documents that users must manually browse through, a question answering system attempts to directly answer natural language questions posed by the user. Although such systems possess language-processing capabilities, they still rely on traditional document retrieval techniques to generate an initial candidate set of documents. In this article, the authors argue that document retrieval for question answering represents a task different from retrieving documents in response to more general retrospective information needs. Thus, to guide future system development, specialized question answering test collections must be constructed. They show that the current evaluation resources have major shortcomings; to remedy the situation, they have manually created a small, reusable question answering test collection for research purposes. In this article they describe their methodology for building this test collection and discuss issues they encountered regarding the notion of "answer correctness."
引用
收藏
页码:851 / 861
页数:11
相关论文
共 50 条
[1]  
[Anonymous], P 9 TEXT RETR C TREC
[2]  
[Anonymous], 2004, P INF RETR QUEST ANS
[3]  
[Anonymous], P 16 ANN INT ACM SIG
[4]  
[Anonymous], P HUM LANG TECHN C S
[5]   Users' criteria for relevance evaluation: A cross-situational comparison [J].
Barry, CL ;
Schamber, L .
INFORMATION PROCESSING & MANAGEMENT, 1998, 34 (2-3) :219-236
[6]  
BILOTTI M, 2004, THESIS MIT CAMBRIDGE
[7]  
Brill E., 2001, P TREC GAITNH MD US, VVolume 56, P90
[8]  
Buckley C., 2004, Proceedings of Sheffield SIGIR 2004. The Twenty-Seventh Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, P25, DOI 10.1145/1008992.1009000
[9]  
Buckley C., 2000, P 23 ANN INT ACM SIG, P33, DOI DOI 10.1145/345508.345543
[10]  
Cieri C, 2002, KLUW S INF, V12, P33