Compact Deep Neural Networks with l1,1 and l1,2 Regularization

被引:0
|
作者
Ma, Rongrong [1 ]
Niu, Lingfeng [2 ]
机构
[1] Univ Chinese Acad Sci, Sch Math Sci, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Econ & Management, Beijing, Peoples R China
来源
2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW) | 2018年
基金
中国国家自然科学基金;
关键词
deep neural networks; sparse regularizer; l(1,1); l(1,2);
D O I
10.1109/ICDMW.2018.00178
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks have demonstrated its superiority in many fields. Its excellent performance relys on quite a lot of parameters used in the network, resulting in a series of problems, including memory and computation requirement and overfitting, which seriously impede the application of deep neural networks in many assignments in practice. A considerable number of model compression methods have been proposed in deep neural networks to reduce the number of parameters used in networks, among which there is one kind of methods persuing sparsity in deep neural networks. In this paper, we propose to combine l(1,1) and l(1,2) norm together as the regularization term to regularize the objective function of the network. We introduce group and l(1,1) can zero out weights in both intergroup and intra-group level. l(1,2) regularizer can obtain intragroup level sparsity and cause even weights among groups. We adopt proximal gradient descent to solve the objective function regularized by our combined regularization. Experimental results demonstrate the effectiveness of the proposed regularizer when comparing it with other baseline regularizers.
引用
收藏
页码:1248 / 1254
页数:7
相关论文
共 11 条
  • [11] S1 and S2 Heart Sound Recognition Using Deep Neural Networks
    Chen, Tien-En
    Yang, Shih-I
    Ho, Li-Ting
    Tsai, Kun-Hsi
    Chen, Yu-Hsuan
    Chang, Yun-Fan
    Lai, Ying-Hui
    Wang, Syu-Siang
    Tsao, Yu
    Wu, Chau-Chung
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2017, 64 (02) : 372 - 380