共 11 条
Compact Deep Neural Networks with l1,1 and l1,2 Regularization
被引:0
|作者:
Ma, Rongrong
[1
]
Niu, Lingfeng
[2
]
机构:
[1] Univ Chinese Acad Sci, Sch Math Sci, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Econ & Management, Beijing, Peoples R China
来源:
2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW)
|
2018年
基金:
中国国家自然科学基金;
关键词:
deep neural networks;
sparse regularizer;
l(1,1);
l(1,2);
D O I:
10.1109/ICDMW.2018.00178
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Deep neural networks have demonstrated its superiority in many fields. Its excellent performance relys on quite a lot of parameters used in the network, resulting in a series of problems, including memory and computation requirement and overfitting, which seriously impede the application of deep neural networks in many assignments in practice. A considerable number of model compression methods have been proposed in deep neural networks to reduce the number of parameters used in networks, among which there is one kind of methods persuing sparsity in deep neural networks. In this paper, we propose to combine l(1,1) and l(1,2) norm together as the regularization term to regularize the objective function of the network. We introduce group and l(1,1) can zero out weights in both intergroup and intra-group level. l(1,2) regularizer can obtain intragroup level sparsity and cause even weights among groups. We adopt proximal gradient descent to solve the objective function regularized by our combined regularization. Experimental results demonstrate the effectiveness of the proposed regularizer when comparing it with other baseline regularizers.
引用
收藏
页码:1248 / 1254
页数:7
相关论文