Adversarial Multimodal Representation Learning for Click-Through Rate Prediction

被引:32
作者
Li, Xiang [1 ,2 ]
Wang, Chao [1 ,2 ]
Tan, Jiwei [1 ,2 ]
Zeng, Xiaoyi [1 ,2 ]
Ou, Dan [1 ,2 ]
Zheng, Bo [1 ,2 ]
机构
[1] Alibaba Grp, Hangzhou, Peoples R China
[2] Alibaba Grp, Beijing, Peoples R China
来源
WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020) | 2020年
关键词
multimodal learning; adversarial learning; recurrent neural network; attention; representation learning; e-commerce search;
D O I
10.1145/3366423.3380163
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
For better user experience and business effectiveness, Click-Through Rate (CTR) prediction has been one of the most important tasks in E-commerce. Although extensive CTR prediction models have been proposed, learning good representation of items from multimodal features is still less investigated, considering an item in E-commerce usually contains multiple heterogeneous modalities. Previous works either concatenate the multiple modality features, that is equivalent to giving a fixed importance weight to each modality; or learn dynamic weights of different modalities for different items through technique like attention mechanism. However, a problem is that there usually exists common redundant information across multiple modalities. The dynamic weights of different modalities computed by using the redundant information may not correctly reflect the different importance of each modality. To address this, we explore the complementarity and redundancy of modalities by considering modality-specific and modality-invariant features differently. We propose a novel Multimodal Adversarial Representation Network (MARN) for the CTR prediction task. A multimodal attention network first calculates the weights of multiple modalities for each item according to its modality-specific features. Then a multimodal adversarial network learns modality-invariant representations where a double-discriminators strategy is introduced. Finally, we achieve the multimodal item representations by combining both modality-specific and modality-invariant representations. We conduct extensive experiments on both public and industrial datasets, and the proposed method consistently achieves remarkable improvements to the state-of-the-art methods. Moreover, the approach has been deployed in an operational E-commerce system and online A/B testing further demonstrates the effectiveness.
引用
收藏
页码:827 / 836
页数:10
相关论文
共 50 条
  • [41] Multimodal Fusion Representation Learning Based on Differential Privacy
    Cai, Chaoxin
    Sang, Yingpeng
    Huang, Jinghao
    Zhang, Maliang
    Li, Weizheng
    PARALLEL AND DISTRIBUTED COMPUTING, APPLICATIONS AND TECHNOLOGIES, PDCAT 2021, 2022, 13148 : 548 - 559
  • [42] Multimodal Cardiac Segmentation Using Disentangled Representation Learning
    Chartsias, Agisilaos
    Papanastasiou, Giorgos
    Wang, Chengjia
    Stirrat, Colin
    Semple, Scott
    Newby, David
    Dharmakumar, Rohan
    Tsaftaris, Sotirios A.
    STATISTICAL ATLASES AND COMPUTATIONAL MODELS OF THE HEART: MULTI-SEQUENCE CMR SEGMENTATION, CRT-EPIGGY AND LV FULL QUANTIFICATION CHALLENGES, 2020, 12009 : 128 - 137
  • [43] Knowledge Base Completion Based on Multimodal Representation Learning
    Wang J.
    Su H.
    Lai X.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2021, 34 (01): : 33 - 43
  • [44] Multimodal Data Enhanced Representation Learning for Knowledge Graphs
    Wang, Zikang
    Li, Linjing
    Li, Qiudan
    Zeng, Daniel
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [45] Multimodal representation models for prediction and control from partial information
    Zambelli, Martina
    Cully, Antoine
    Demiris, Yiannis
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2020, 123 (123)
  • [46] Feature Equilibrium: An Adversarial Training Method to Improve Representation Learning
    Liu, Minghui
    Yang, Meiyi
    Deng, Jiali
    Cheng, Xuan
    Xie, Tianshu
    Deng, Pan
    Gong, Haigang
    Liu, Ming
    Wang, Xiaomin
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2023, 16 (01)
  • [47] Multimodal salient object detection via adversarial learning with collaborative generator
    Tu, Zhengzheng
    Yang, Wenfang
    Wang, Kunpeng
    Hussain, Amir
    Luo, Bin
    Li, Chenglong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 119
  • [48] Feature Equilibrium: An Adversarial Training Method to Improve Representation Learning
    Minghui Liu
    Meiyi Yang
    Jiali Deng
    Xuan Cheng
    Tianshu Xie
    Pan Deng
    Haigang Gong
    Ming Liu
    Xiaomin Wang
    International Journal of Computational Intelligence Systems, 16
  • [49] Multimodal Adversarial Learning Based Unsupervised Time Series Anomaly Detection
    Huang X.
    Zhang F.
    Fan H.
    Xi L.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (08): : 1655 - 1667
  • [50] Adversarial Representation Learning for Intelligent Condition Monitoring of Complex Machinery
    Sun, Shilin
    Wang, Tianyang
    Yang, Hongxing
    Chu, Fulei
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (05) : 5255 - 5265