Compact Deep Neural Networks with l1,1 and l1,2 Regularization

被引:0
|
作者
Ma, Rongrong [1 ]
Niu, Lingfeng [2 ]
机构
[1] Univ Chinese Acad Sci, Sch Math Sci, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Econ & Management, Beijing, Peoples R China
来源
2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW) | 2018年
基金
中国国家自然科学基金;
关键词
deep neural networks; sparse regularizer; l(1,1); l(1,2);
D O I
10.1109/ICDMW.2018.00178
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks have demonstrated its superiority in many fields. Its excellent performance relys on quite a lot of parameters used in the network, resulting in a series of problems, including memory and computation requirement and overfitting, which seriously impede the application of deep neural networks in many assignments in practice. A considerable number of model compression methods have been proposed in deep neural networks to reduce the number of parameters used in networks, among which there is one kind of methods persuing sparsity in deep neural networks. In this paper, we propose to combine l(1,1) and l(1,2) norm together as the regularization term to regularize the objective function of the network. We introduce group and l(1,1) can zero out weights in both intergroup and intra-group level. l(1,2) regularizer can obtain intragroup level sparsity and cause even weights among groups. We adopt proximal gradient descent to solve the objective function regularized by our combined regularization. Experimental results demonstrate the effectiveness of the proposed regularizer when comparing it with other baseline regularizers.
引用
收藏
页码:1248 / 1254
页数:7
相关论文
共 11 条
  • [1] Transformed l1 regularization for learning sparse deep neural networks
    Ma, Rongrong
    Miao, Jianyu
    Niu, Lingfeng
    Zhang, Peng
    NEURAL NETWORKS, 2019, 119 : 286 - 298
  • [2] Training Compact DNNs with l1/2 Regularization
    Tang, Anda
    Niu, Lingfeng
    Miao, Jianyu
    Zhang, Peng
    PATTERN RECOGNITION, 2023, 136
  • [3] Towards l1 Regularization for Deep Neural Networks: Model Sparsity Versus Task Difficulty
    Shen, Ta-Chun
    Yang, Chun-Pai
    Yen, Ian En-Hsu
    Lin, Shou-De
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2022, : 126 - 134
  • [4] SPARSE DEEP NEURAL NETWORKS USING L1,∞-WEIGHT NORMALIZATION
    Wen, Ming
    Xu, Yixi
    Zheng, Yunling
    Yang, Zhouwang
    Wang, Xiao
    STATISTICA SINICA, 2021, 31 (03) : 1397 - 1414
  • [5] Intelligent fault diagnosis for rotating machinery using L1/2-SF under variable rotational speed
    Wang, Jinrui
    Ji, Shanshan
    Han, Baokun
    Bao, Huaiqian
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2021, 235 (05) : 1409 - 1422
  • [6] Mispronunciation Detection and Diagnosis in L2 English Speech Using Multidistribution Deep Neural Networks
    Li, Kun
    Qian, Xiaojun
    Meng, Helen
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2017, 25 (01) : 193 - 207
  • [7] Intonation classification for L2 English speech using multi-distribution deep neural networks
    Li, Kun
    Wu, Xixin
    Meng, Helen
    COMPUTER SPEECH AND LANGUAGE, 2017, 43 : 18 - 33
  • [8] Mispronunciation Detection and Diagnosis in L2 English Speech Using Multi-Distribution Deep Neural Networks
    Li, Kun
    Meng, Helen
    2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 255 - 259
  • [9] Automatic lexical stress and pitch accent detection for L2 English speech using multi-distribution deep neural networks
    Li, Kun
    Mao, Shaoguang
    Li, Xu
    Wu, Zhiyong
    Meng, Helen
    SPEECH COMMUNICATION, 2018, 96 : 28 - 36
  • [10] Neural Networks or Linguistic Features? - Comparing Different Machine-Learning Approaches for Automated Assessment of Text Quality Traits Among L1-and L2-Learners' Argumentative Essays
    Lohmann, Julian F.
    Junge, Fynn
    Moeller, Jens
    Fleckenstein, Johanna
    Trub, Ruth
    Keller, Stefan
    Jansen, Thorben
    Horbach, Andrea
    INTERNATIONAL JOURNAL OF ARTIFICIAL INTELLIGENCE IN EDUCATION, 2024,