A new interest extraction method based on multi-head attention mechanism for CTR prediction

被引:4
|
作者
Yang, Haifeng [1 ]
Yao, Linjing [1 ]
Cai, Jianghui [1 ,2 ]
Wang, Yupeng [1 ]
Zhao, Xujun [1 ]
机构
[1] Taiyuan Univ Sci & Technol, Sch Comp Sci & Technol, Waliu Rd, Taiyuan 030024, Peoples R China
[2] North Univ China, Sch Comp Sci & Technol, Xueyuan Rd, Taiyuan 030051, Peoples R China
基金
中国国家自然科学基金;
关键词
Recommendation system; Multi-head attention; Feature interaction; Click-through rate prediction;
D O I
10.1007/s10115-023-01867-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Click-through rate (CTR) prediction plays a vital role in recommendation systems. Most models pay little attention to the relationship between target items in the user behavior sequence. The attention units used in these models cannot fully capture the context information, which can be used to reflect the variations of user interests. To address these problems, we propose a new model named interest extraction method based on multi-head attention mechanism (IEN) for CTR prediction. Specifically, we design an interest extraction module, which consists of two sub-modules: the item representation module (IRM) and the context-item interaction module (CIM). In IRM, we learn the relationship between target items in the user behavior sequence by a multi-head attention mechanism. Then, the user representation is gained by integrating the refined item representation and position information. At last, the correlation between the user and the target item is used to reflect user interests. In CIM, the context information has valuable temporal features which can reflect the variations of user interests. Therefore, user interests can be further acquired through the feature interaction between the context and the target item. After that, the learned relevance and the feature interaction are fed to the multi-layer perceptron (MLP) for prediction. Besides, experiments on four Amazon datasets were conducted to evaluate the effectiveness of our method in capturing user interests. The experimental results show that our proposed method outperforms state-of-the-art methods in terms of AUC and RI in the CTR prediction task.
引用
收藏
页码:3337 / 3352
页数:16
相关论文
共 50 条
  • [31] Software and Hardware Fusion Multi-Head Attention
    Hu, Wei
    Xu, Dian
    Liu, Fang
    Fan, Zimeng
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 644 - 655
  • [32] Classification of Heads in Multi-head Attention Mechanisms
    Huang, Feihu
    Jiang, Min
    Liu, Fang
    Xu, Dian
    Fan, Zimeng
    Wang, Yonghao
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2022, PT III, 2022, 13370 : 681 - 692
  • [33] Diversifying Multi-Head Attention in the Transformer Model
    Ampazis, Nicholas
    Sakketou, Flora
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (04): : 2618 - 2638
  • [34] Multi-head Self-attention Recommendation Model based on Feature Interaction Enhancement
    Yin, Yunfei
    Huang, Caihao
    Sun, Jingqin
    Huang, Faliang
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 1740 - 1745
  • [35] Improving Multi-head Attention with Capsule Networks
    Gu, Shuhao
    Feng, Yang
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 314 - 326
  • [36] Optimizing the Online Learners' Verbal Intention Classification Efficiency Based on the Multi-Head Attention Mechanism Algorithm
    Zheng, Yangfeng
    Shao, Zheng
    Gao, Zhanghao
    Deng, Mingming
    Zhai, Xuesong
    INTERNATIONAL JOURNAL OF FOUNDATIONS OF COMPUTER SCIENCE, 2022, 33 (06N07) : 717 - 733
  • [37] Improved Convolutional Neural Network Based on Multi-head Attention Mechanism for Industrial Process Fault Classification
    Cui, Wenzhi
    Deng, Xiaogang
    Zhang, Zheng
    PROCEEDINGS OF 2020 IEEE 9TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS'20), 2020, : 918 - 922
  • [38] RMAN: Relational multi-head attention neural network for joint extraction of entities and relations
    Taiqu Lai
    Lianglun Cheng
    Depei Wang
    Haiming Ye
    Weiwen Zhang
    Applied Intelligence, 2022, 52 : 3132 - 3142
  • [39] Text classification model based on multi-head attention capsule neworks
    Jia X.
    Wang L.
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2020, 60 (05): : 415 - 421
  • [40] A Novel Source Code Representation Approach Based on Multi-Head Attention
    Xiao, Lei
    Zhong, Hao
    Liu, Jianjian
    Zhang, Kaiyu
    Xu, Qizhen
    Chang, Le
    ELECTRONICS, 2024, 13 (11)