Evaluating the effectiveness of information retrieval systems using effort-based relevance judgment

被引:1
作者
Rajagopal, Prabha [1 ]
Ravana, Sri Devi [1 ]
Koh, Yun Sing [2 ]
Balakrishnan, Vimala [1 ]
机构
[1] Univ Malaya, Kuala Lumpur, Malaysia
[2] Univ Auckland, Auckland, New Zealand
关键词
Information system; Information retrieval; TREC; Large-scale experimentation; Relevance judgments; System-oriented evaluation;
D O I
10.1108/AJIM-04-2018-0086
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Purpose The effort in addition to relevance is a major factor for satisfaction and utility of the document to the actual user. The purpose of this paper is to propose a method in generating relevance judgments that incorporate effort without human judges' involvement. Then the study determines the variation in system rankings due to low effort relevance judgment in evaluating retrieval systems at different depth of evaluation. Design/methodology/approach Effort-based relevance judgments are generated using a proposed boxplot approach for simple document features, HTML features and readability features. The boxplot approach is a simple yet repeatable approach in classifying documents' effort while ensuring outlier scores do not skew the grading of the entire set of documents. Findings The retrieval systems evaluation using low effort relevance judgments has a stronger influence on shallow depth of evaluation compared to deeper depth. It is proved that difference in the system rankings is due to low effort documents and not the number of relevant documents. Originality/value Hence, it is crucial to evaluate retrieval systems at shallow depth using low effort relevance judgments.
引用
收藏
页码:2 / 17
页数:16
相关论文
共 20 条
[1]   A Review of Factors Influencing User Satisfaction in Information Retrieval [J].
Al-Maskari, Azzah ;
Sanderson, Mark .
JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY, 2010, 61 (05) :859-868
[2]  
Carterette B, 2010, SIGIR 2010: PROCEEDINGS OF THE 33RD ANNUAL INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH DEVELOPMENT IN INFORMATION RETRIEVAL, P539
[3]  
Chandar P, 2013, SIGIR'13: THE PROCEEDINGS OF THE 36TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH & DEVELOPMENT IN INFORMATION RETRIEVAL, P745
[4]  
Collins-Thompson K., 2011, SIGIR 2011 Workshop on Enriching Information Retrieval (ENIR 2011), P1
[5]  
DuBay W.H., 2004, PRINCIPLES READABILI, DOI [10. 1080/07060668509501661, DOI 10.1080/07060668509501661]
[6]   A Few Good Topics: Experiments in Topic Set Reduction for Retrieval Evaluation [J].
Guiver, John ;
Mizzaro, Stefano ;
Robertson, Stephen .
ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2009, 27 (04)
[7]  
Gustafsdottir G.U., 2017, Readability tests - Siteimprove Help Center
[8]  
Hersh W., 2000, SIGIR Forum, V34, P17
[9]  
Jones T., 2014, Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management - CIKM'14, P1843, DOI DOI 10.1145/2661829.2661945IVER
[10]  
Moffat A., 2012, P 17 AUSTR DOC COMP, P47, DOI [10.1145/2407085.2407092, DOI 10.1145/2407085.2407092, 10.1145/2407085.2407092ty, DOI 10.1145/2407085.2407092TY]