Deep and Shallow Model for Insurance Churn Prediction Service

被引:17
作者
Zhang, Rong [1 ]
Li, Weiping [1 ]
Mo, Tong [1 ]
Tan, Wei [2 ]
机构
[1] Peking Univ, Sch Software & Microelect, Beijing, Peoples R China
[2] IBM Corp, TJ Watson Res Ctr, Yorktown Hts, NY USA
来源
2017 IEEE INTERNATIONAL CONFERENCE ON SERVICES COMPUTING (SCC) | 2017年
基金
中国国家自然科学基金;
关键词
Deep & Shallow Model; Joint Training; Machine Learning; Deep Learning; Insurance Churn Prediction Service;
D O I
10.1109/SCC.2017.51
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Churn prediction is very important to the insurance industry. Therefore, there is a big value to investigate how to improve its performance. More importantly, a good model can be used by a common service provider and benefit many companies. State-of-the-art methods either use 1) shallow models such as logistic regression, with sophisticated feature engineering, or 2) deep models that learn features and classification models simultaneously. In terms of performance, shallow models can memorize better while deep models can generalize better but may under-generalize with insufficient data. Therefore, we propose a combined Deep & Shallow model (DSM) to take the strengths of both memorization and generalization in one model by jointly training shallow models and deep models. The experiment results show that for insurance churn prediction, joint training can significantly improve the performance and the DSM earns better performance than both shallow-only and deep-only models. In our real-life dataset, the DSM performs better than CNN, LSTM, Stochastic Gradient Descent, Linear Discriminant Analysis, Quadratic Discriminant Analysis, Gaussian Naive Bayes, AdaBoost, Random Forest, and Gradient Tree Boosting. In addition, the DSM can also be applied to other prediction services.
引用
收藏
页码:346 / 353
页数:8
相关论文
共 32 条
[1]  
[Anonymous], INT C ADV INF COMM T
[2]  
[Anonymous], PROC CVPR IEEE
[3]  
[Anonymous], NEUROCOMPUTING
[4]  
[Anonymous], 2014, C EMPIRICAL METHODS
[5]  
[Anonymous], 2015, IEEE I CONF COMP VIS, DOI DOI 10.1109/ICCV.2015.123
[6]  
[Anonymous], P 1 WORKSH DEEP LEAR, DOI [10.1145/2988450.2988454, DOI 10.1145/2988450.2988454]
[7]  
[Anonymous], 2012, MACHINE LEARNING PRO
[8]  
Arjovsky M., 2017, ARXIV170107875
[9]  
Batista GE., 2003, Balancing Training Data for Automated Annotation of Keywords: A Case Study
[10]  
Batista GE., 2004, ACM SIGKDD EXPL NEWS, V6, P20, DOI DOI 10.1145/1007730.1007735