Diversifying Search Results using Self-Attention Network

被引:26
作者
Qin, Xubo [2 ]
Dou, Zhicheng [1 ]
Wen, Ji-Rong [3 ,4 ]
机构
[1] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
[2] Renmin Univ China, Sch Informat, Beijing, Peoples R China
[3] Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
[4] MOE, Key Lab Data Engn & Knowledge Engn, Beijing, Peoples R China
来源
CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT | 2020年
关键词
Search Result Diversification; Self Attention;
D O I
10.1145/3340531.3411914
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Search results returned by search engines need to be diversified in order to satisfy different information needs of different users. Several supervised learning models have been proposed for diversifying search results in recent years. Most of the existing supervised methods greedily compare each candidate document with the selected document sequence and select the next local optimal document. However, the information utility of each candidate document is not independent with each other, and research has shown that the selection of a candidate document will affect the utilities of other candidate documents. As a result, the local optimal document rankings will not lead to the global optimal rankings. In this paper, we propose a new supervised diversification framework to address this issue. Based on a self-attention encoder-decoder structure, the model can take the whole candidate document sequence as input, and simultaneously leverage both the novelty and the subtopic coverage of the candidate documents. We call this framework Diversity Encoder with Self-Attention (DESA). Comparing with existing supervised methods, this framework can model the interactions between all candidate documents and return their diversification scores based on the whole candidate document sequence. Experimental results show that our proposed framework outperforms existing methods. These results confirm the effectiveness of modeling all the candidate documents for the overall novelty and subtopic coverage globally, instead of comparing every single candidate document with the selected sequence document selection.
引用
收藏
页码:1265 / 1274
页数:10
相关论文
共 32 条
[1]  
Agrawal Rakesh, 2009, WSDM, P5, DOI 10.1145/1498759.1498766
[2]  
Ba JL., 2016, ARXIV
[3]  
Baeza-Yates Ricardo, CURR TRENDS DAT TECH, P588
[4]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[5]  
Carbonell J., 1998, Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, P335, DOI 10.1145/290941.291025
[6]  
Chapelle O, 2009, P 18 ACM C INF KNOWL, P621, DOI DOI 10.1145/1645953.1646033
[7]  
ChengXiang Zhai, 2015, ACM SIGIR Forum, V49, P2
[8]  
Clarke Charles L, 2009, Technical Report
[9]  
Clarke Charles L. A., 2008, P 31 ANN INT ACM SIG, P659, DOI [DOI 10.1145/1390334.1390446, 10.1145/1390334.1390446]
[10]  
Dang V, 2012, SIGIR 2012: PROCEEDINGS OF THE 35TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, P65, DOI 10.1145/2348283.2348296