A new interest extraction method based on multi-head attention mechanism for CTR prediction

被引:4
|
作者
Yang, Haifeng [1 ]
Yao, Linjing [1 ]
Cai, Jianghui [1 ,2 ]
Wang, Yupeng [1 ]
Zhao, Xujun [1 ]
机构
[1] Taiyuan Univ Sci & Technol, Sch Comp Sci & Technol, Waliu Rd, Taiyuan 030024, Peoples R China
[2] North Univ China, Sch Comp Sci & Technol, Xueyuan Rd, Taiyuan 030051, Peoples R China
基金
中国国家自然科学基金;
关键词
Recommendation system; Multi-head attention; Feature interaction; Click-through rate prediction;
D O I
10.1007/s10115-023-01867-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Click-through rate (CTR) prediction plays a vital role in recommendation systems. Most models pay little attention to the relationship between target items in the user behavior sequence. The attention units used in these models cannot fully capture the context information, which can be used to reflect the variations of user interests. To address these problems, we propose a new model named interest extraction method based on multi-head attention mechanism (IEN) for CTR prediction. Specifically, we design an interest extraction module, which consists of two sub-modules: the item representation module (IRM) and the context-item interaction module (CIM). In IRM, we learn the relationship between target items in the user behavior sequence by a multi-head attention mechanism. Then, the user representation is gained by integrating the refined item representation and position information. At last, the correlation between the user and the target item is used to reflect user interests. In CIM, the context information has valuable temporal features which can reflect the variations of user interests. Therefore, user interests can be further acquired through the feature interaction between the context and the target item. After that, the learned relevance and the feature interaction are fed to the multi-layer perceptron (MLP) for prediction. Besides, experiments on four Amazon datasets were conducted to evaluate the effectiveness of our method in capturing user interests. The experimental results show that our proposed method outperforms state-of-the-art methods in terms of AUC and RI in the CTR prediction task.
引用
收藏
页码:3337 / 3352
页数:16
相关论文
共 50 条
  • [41] RMAN: Relational multi-head attention neural network for joint extraction of entities and relations
    Lai, Taiqu
    Cheng, Lianglun
    Wang, Depei
    Ye, Haiming
    Zhang, Weiwen
    APPLIED INTELLIGENCE, 2022, 52 (03) : 3132 - 3142
  • [42] Duplicate Question Detection based on Neural Networks and Multi-head Attention
    Zhang, Heng
    Chen, Liangyu
    PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2019, : 13 - 18
  • [43] Multi-head Attention and Graph Convolutional Networks with Regularized Dropout for Biomedical Relation Extraction
    Huang, Mian
    Wang, Jian
    Lin, Hongfei
    Yang, Zhihao
    HEALTH INFORMATION PROCESSING, CHIP 2023, 2023, 1993 : 98 - 111
  • [44] Relation Extraction in Biomedical Texts Based on Multi-Head Attention Model With Syntactic Dependency Feature: Modeling Study
    Li, Yongbin
    Hui, Linhu
    Zou, Liping
    Li, Huyang
    Xu, Luo
    Wang, Xiaohua
    Chua, Stephanie
    JMIR MEDICAL INFORMATICS, 2022, 10 (10)
  • [45] Multi-head attention-based intelligent vehicle lane change decision and trajectory prediction model in highways
    Cai, Junyu
    Jiang, Haobin
    Wang, Junyan
    Li, Aoxue
    JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS, 2024,
  • [46] A Graph Neural Network Social Recommendation Algorithm Integrating the Multi-Head Attention Mechanism
    Yi, Huawei
    Liu, Jingtong
    Xu, Wenqian
    Li, Xiaohui
    Qian, Huihui
    ELECTRONICS, 2023, 12 (06)
  • [47] Accurate prediction of drug combination risk levels based on relational graph convolutional network and multi-head attention
    He, Shi-Hui
    Yun, Lijun
    Yi, Hai-Cheng
    JOURNAL OF TRANSLATIONAL MEDICINE, 2024, 22 (01)
  • [48] Acoustic Scene Analysis with Multi-head Attention Networks
    Wang, Weimin
    Wang, Weiran
    Sun, Ming
    Wang, Chao
    INTERSPEECH 2020, 2020, : 1191 - 1195
  • [49] Enhancing Attention Models via Multi-head Collaboration
    Wang, Huadong
    Tu, Mei
    2020 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2020), 2020, : 19 - 23
  • [50] Bidirectional GRU with Multi-Head Attention for Chinese NER
    Yan, Shuo
    Chai, Jianping
    Wu, Liyun
    PROCEEDINGS OF 2020 IEEE 5TH INFORMATION TECHNOLOGY AND MECHATRONICS ENGINEERING CONFERENCE (ITOEC 2020), 2020, : 1160 - 1164