Residual connections improve click-through rate and conversion rate prediction performance

被引:0
作者
Ergun Biçici [1 ]
机构
[1] AI Enablement, Huawei Türkiye R & D Center, Istanbul
关键词
Click-through rate; CTR; CVR; DNN; Neural networks; Residual connections;
D O I
10.1007/s00521-024-10617-0
中图分类号
学科分类号
摘要
The prediction of click-through rate (CTR) and conversion rate (CVR) are crucial tasks in online advertising and recommendation systems. As the learning models become more complex with increasing depth, it has become increasingly challenging to predict CTR and CVR accurately. This paper addresses the challenges associated with the increasing depth in CTR and CVR prediction models by introducing the integration of residual connections into the models. The experiments we conducted involve using five different CTR or CVR prediction models together with residual connections on benchmark datasets from both Avazu and Criteo, and the company dataset. The results demonstrate that residual connections can effectively improve CTR and CVR prediction models, with an increase in AUC by 1.4%, a decrease in loss by 3.5%, and an increase in F1 by 20.2%. The results also show that we can safely increase the depth and the size of the network without a need to optimize or a decrease in the performance. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:10675 / 10688
页数:13
相关论文
共 38 条
  • [1] Szegedy C., Liu W., Jia Y., Sermanet P., Reed S., Anguelov D., Erhan D., Vanhoucke V., Rabinovich A., Going deeper with convolutions, In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)., pp. 1-9, (2015)
  • [2] Simonyan K., Zisserman A., Very deep convolutional networks for large-scale image recognition, . In: Bengio Y, Lecun Y. (Eds.) 3Rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 2015, Conference Track Proceedings, (2015)
  • [3] Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., Kaiser L., Polosukhin I., Attention is all you need, In: Proceedings of the 31St International Conference on Neural Information Processing Systems. NIPS’17., pp. 6000-6010, (2017)
  • [4] Brown T.B., Mann B., Ryder N., Subbiah M., Kaplan J., Dhariwal P., Neelakantan A., Shyam P., Sastry G., Askell A., Agarwal S., Herbert-Voss A., Krueger G., Henighan T., Child R., Ramesh A., Ziegler D.M., Wu J., Winter C., Hesse C., Chen M., Sigler E., Litwin M., Gray S., Chess B., Clark J., Berner C., McCandlish S., Radford A., Sutskever I., Amodei D., Language Models are Few-Shot Learners, (2020)
  • [5] Brown T.B., Mann B., Ryder N., Subbiah M., Kaplan J., Dhariwal P., Neelakantan A., Shyam P., Sastry G., Askell A., Agarwal S., Herbert-Voss A., Krueger G., Henighan T., Child R., Ramesh A., Ziegler D.M., Wu J., Winter C., Hesse C., Chen M., Sigler E., Litwin M., Gray S., Chess B., Clark J., Berner C., McCandlish S., Radford A., Sutskever I., Amodei D., Language models are few-shot learners, . In: Proceedings of the 34Th International Conference on Neural Information Processing Systems. NIPS’20.
  • [6] He K., Zhang X., Ren S., Sun J., Deep residual learning for image recognition, 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp. 770-778, (2016)
  • [7] He K., Zhang X., Ren S., Sun J., Identity mappings in deep residual networks, Computer vision–ECCV 2016, pp. 630-645, (2016)
  • [8] Ramesh A., Pavlov M., Goh G., Gray S., Voss C., Radford A., Chen M., Sutskever I., Zero-shot text-to-image generation, Proceedings of the 38Th International Conference on Machine Learning. Proceedings of Machine Learning Research, PMLR, Virtual Only, 139, pp. 8821-8831, (2021)
  • [9] Radford A., Wu J., Child R., Luan D., Amodei D., Sutskever I., Language Models are Unsupervised Multitask Learners, (2019)
  • [10] Li H., Xu Z., Taylor G., Studer C., Goldstein T., Visualizing the loss landscape of neural nets, Proceedings of the 32St International Conference on Neural Information Processing Systems. NIPS’18, 31, (2018)