Click-Through Rate Prediction of Multi-Head Self-Attention in Hyperbolic Space

被引:0
作者
Han Y.-L. [1 ]
Wang X.-Y. [1 ]
机构
[1] School of Computer Science and Technology, Harbin University of Science and Technology, Harbin
来源
Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications | 2021年 / 44卷 / 05期
关键词
Hyperbolic space; Lorentzian distance; Multi-head self-attention;
D O I
10.13190/j.jbupt.2021-017
中图分类号
学科分类号
摘要
In recommendation systems, understanding the complex functional interactions behind user behaviors is crucial to predict the clicking probability of users on advertisements or commodities. Efforts have been made to find low-dimensional representations and meaningful combinations of sparse and high-dimensional original features. Among them, the deep & cross network can explicitly cross features at each layer. However it treats all crossing features "equally" and does not consider the influence of different features on the results, which may eliminate some useful information. Therefore, a prediction model of click-through rate of multi-head self-attention neural network in hyperbolic space is proposed. In hyperbolic space, the model uses Lorentzian distance instead of inner product, to measure the similarity and correlation between features, which can avoid dimension disaster. Experimental results show that the model is superior to the deep & cross network on predicting click-through rate data sets in terms of accuracy. © 2021, Editorial Department of Journal of Beijing University of Posts and Telecommunications. All right reserved.
引用
收藏
页码:127 / 132
页数:5
相关论文
共 17 条
  • [1] He Xiangnan, Chua Tat-Seng, Neural factorization machines for sparse predictive analytics, SIGIR'17: The 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 355-364, (2017)
  • [2] Cheng Heng-Tze, Koc L, Harmsen J, Et al., Wide & deep learning for recommender systems, DLRS 2016: The 1st Workshop on Deep Learning for Recommendder Systems, pp. 7-10, (2016)
  • [3] Rendle S., Factorization machines, 2010 IEEE International Conference on Data Mining, pp. 995-1000, (2010)
  • [4] Beutel A, Covington P, Jain S, Et al., Latent cross: making use of context in recurrent recommender systems, WSDM'18: The Eleventh ACM International Conference on Web Search and Data Mining, pp. 46-54, (2018)
  • [5] Vaswani A, Shazeer N, Parmar N, Et al., Attention is all you need, NIPS'17: The 31st International Confe-rence on Neural Information Processing Systems, pp. 5998-6008, (2017)
  • [6] Hsieh C, Yang Longqi, Yin Cui, Et al., Collaborative metric learning, WWW'17: The 26th International Conference on World Wide Web, pp. 193-201, (2017)
  • [7] Tran L, Tay Y, Zhang Shuai, Et al., HyperML: a boosting metric learning approach in hyperbolic space for recommender systems, WSDM'20: The 13th International Conference on Web Search and Data Mining, pp. 609-617, (2020)
  • [8] Miller A, Fisch A, Dodge J, Et al., Key-value memory networks for directly reading documents, EMNLP 2016: Conference on Empirical Methods in Natural Language Processing, pp. 1400-1409, (2016)
  • [9] Criteo_dataset
  • [10] KDD cup 2012, track 2