Click-through rate prediction using transfer learning with fine-tuned parameters

被引:12
作者
Yang, Xiangli [1 ]
Liu, Qing [2 ]
Su, Rong [2 ]
Tang, Ruiming [2 ]
Liu, Zhirong [2 ]
He, Xiuqiang [2 ]
Yang, Jianxi [1 ]
机构
[1] Chongqing Jiaotong Univ, Sch Informat Sci & Engn, Chongqing, Peoples R China
[2] Huawei Technol Co Ltd, Noahs Ark Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Transfer Learning; Recommender Systems; CTR Prediction; Automatic Fine-tuning;
D O I
10.1016/j.ins.2022.08.009
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In real business platforms, recommendation systems usually need to predict the CTR of multiple business. Since different scenarios may have common feature interactions, knowledge transferring based methods are often used by re-optimizing the pre-trained CTR model from source scenarios to a new target domain. In addition to knowledge transfer, it is noteworthy that generalizing target domain data outside of the CTR model accurately is also important when re-training all of the fine-tuned parameters. Generally, the pretrained model trained on large source domains can represent the characteristics of different instances and capture typical feature interactions. It would be useful to directly reuse fine-tuned parameters from source domains to serve the target domain. However, different instances of the target domain may need different amounts of source information to finetune the model parameters, and these decisions of freezing or re-optimizing model parameters, which highly depend on the fine-tuned model and target instances, may require much manual effort. In this paper, we propose an end-to-end transfer learning framework with fine-tuned parameters for CTR prediction, called Automatic Fine-Tuning (AutoFT). The principal component of AutoFT is a set of learnable transfer policies that independently determine how the specific instance-based fine-tuning policies should be trained, which decide the routing in the embedding representations and the high-order feature representations layer by layer in deep CTR model. Extensive tests on two benchmarks and one real commercial recommender system deployed in Huawei's App Store show that AutoFT can greatly increase CTR prediction performance when compared to current transferring methodologies.
引用
收藏
页码:188 / 200
页数:13
相关论文
共 39 条
  • [1] Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
  • [2] Domain Adaptation in Display Advertising: An Application for Partner Cold-Start
    Aggarwal, Karan
    Yadav, Pranjul
    Keerthi, S. Sathiya
    [J]. RECSYS 2019: 13TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2019, : 178 - 186
  • [3] Cao B, 2010, AAAI CONF ARTIF INTE, P407
  • [4] Cheng H.-. T., 2016, P 1 WORKSH DEEP LEAR, P7, DOI DOI 10.1145/2988450.2988454
  • [5] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [6] Ganin Y, 2015, PR MACH LEARN RES, V37, P1180
  • [7] Cross-domain Recommendation Without Sharing User-relevant Data
    Gao, Chen
    Chen, Xiangning
    Feng, Fuli
    Zhao, Kai
    He, Xiangnan
    Li, Yong
    Jin, Depeng
    [J]. WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 491 - 502
  • [8] Guo HF, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1725
  • [9] SpotTune: Transfer Learning through Adaptive Fine-tuning
    Guo, Yunhui
    Shi, Honghui
    Kumar, Abhishek
    Grauman, Kristen
    Rosing, Tajana
    Feris, Rogerio
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4800 - 4809
  • [10] The MovieLens Datasets: History and Context
    Harper, F. Maxwell
    Konstan, Joseph A.
    [J]. ACM TRANSACTIONS ON INTERACTIVE INTELLIGENT SYSTEMS, 2016, 5 (04)