A Lightweight Knowledge Distillation and Feature Compression Model for User Click-Through Rates Prediction in Edge Computing Scenarios

被引:3
作者
Yang, Bin [1 ,2 ]
Zhou, Jiawei [4 ]
Zhang, Shihao [4 ]
Xing, Ying [4 ]
Jiang, Weiwei [5 ]
Xu, Lexi [3 ]
机构
[1] China Unicom Res Inst, Graph Neural Network, Beijing 100048, Peoples R China
[2] China Unicom Res Inst, Artificial Intelligence Team, Beijing 100048, Peoples R China
[3] China Unicom Res Inst, Network Intelligent Operat Res Ctr, Beijing, Peoples R China
[4] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
[5] Beijing Univ Posts & Telecommun, Sch Informat & Commun Engn, Key Lab Universal Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
关键词
Predictive models; Computational modeling; Data models; Vectors; Analytical models; Accuracy; Performance evaluation; Training; Recommender systems; Frequency modulation; Click-through rate (CTR) prediction; edge learning; feature compression; Internet of Things (IoT) systems; knowledge distillation (KD);
D O I
10.1109/JIOT.2024.3446640
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Along with the development of Internet of Things systems, numerous edge intelligent devices can obtain a large amount of user data, and the analysis of this user data can be applied to business scenarios, such as user click-through rate (CTR) prediction. In the recommendation, advertising and other scenarios, the users at the edge have high response requirements for CTR prediction model training and inference. In the current edge scenario of CTR prediction, there are problems of overly complex model structure and highly sparse original features, which makes it difficult to deploy CTR prediction models at the edge. Therefore, we propose KD-based graph attention FI model (KD-GAFIM), a lightweight recommendation algorithm that combines graph neural networks (GNNs) with knowledge distillation (KD). The approach uses graph attention networks to flexibly capture feature dependencies in a way that maintains a small model size while augmenting the feature vector with feature dependencies. And by sharing the embedding layer of the teacher model, KD-GAFIM improves the efficiency of user CTR prediction. On top of that, we also propose a feature compression strategy guided by model interpretability, which identifies high-contributing features for inference and model refinement based on their performance in model interpretability. This strategy improves efficiency, making KD-GAFIM suitable for training and inference on edge devices. We conducted extensive experiments on multiple datasets. The experimental results show that KD-GAFIM outperforms various state-of-the-art CTR prediction models, demonstrating that GNN-based KD models can improve model performance while reducing model size and feature dimensionality, and have significant potential for application at the edge.
引用
收藏
页码:2295 / 2308
页数:14
相关论文
共 50 条
[1]  
Ali Z., 2015, International Journal of Computer Applications, V128, P975
[2]   The Hidden Cost of the Edge: A Performance Comparison of Edge and Cloud Latencies [J].
Ali-Eldin, Ahmed ;
Wang, Bin ;
Shenoy, Prashant .
SC21: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2021,
[3]  
Asif U, 2020, Arxiv, DOI arXiv:1909.08097
[4]  
Banjanovic-Mehmedovic L., Edge AI: Reshaping the future of edge computing with artificial intelligence
[5]  
Callara M, 2018, PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON APPLIED SMART SYSTEMS (ICASS)
[6]   Deep Learning With Edge Computing: A Review [J].
Chen, Jiasi ;
Ran, Xukan .
PROCEEDINGS OF THE IEEE, 2019, 107 (08) :1655-1674
[7]   Distilling Knowledge via Knowledge Review [J].
Chen, Pengguang ;
Liu, Shu ;
Zhao, Hengshuang ;
Jia, Jiaya .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :5006-5015
[8]   DCAP: Deep Cross Attentional Product Network for User Response Prediction [J].
Chen, Zekai ;
Zhong, Fangtian ;
Chen, Zhumin ;
Zhang, Xiao ;
Pless, Robert ;
Cheng, Xiuzhen .
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, :221-230
[9]  
Cheng Heng-Tze, 2016, P 1 WORKSHOP DEEP LE, P7
[10]  
Chu Y., 2021, arXiv