A Group Regularization Framework of Convolutional Neural Networks Based on the Impact of Lp Regularizers on Magnitude

被引:0
作者
Li, Feng [1 ]
Hu, Yaokai [1 ]
Zhang, Huisheng [1 ]
Deng, Ansheng [2 ]
Zurada, Jacek M. [3 ,4 ]
机构
[1] Dalian Maritime Univ, Sch Sci, Dalian 116026, Peoples R China
[2] Dalian Maritime Univ, Informat Sci & Technol Coll, Dalian 116026, Peoples R China
[3] Univ Louisville, Dept Elect & Comp Engn, Louisville, KY 40292 USA
[4] Univ Social Sci, PL-90113 Lodz, Poland
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2024年 / 54卷 / 12期
基金
中国博士后科学基金;
关键词
Convolutional neural networks (CNNs); filter pruning; group regularization; HGL(1,2)&L-1/2; magnitude; L-1/2; REGULARIZATION; CONVERGENCE;
D O I
10.1109/TSMC.2024.3453549
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Group regularization is commonly employed in network pruning to achieve structured model compression. However, the rationale behind existing studies on group regularization predominantly hinges on the sparsity capabilities of L-p regularizers. This singular focus may lead to erroneous interpretations. In response to these limitations, this article proposes a novel framework for evaluating the penalization efficacy of group regularization methods by analyzing the impact of L-p regularizers on weight magnitudes and weight group magnitudes. Within this framework, we demonstrate that L-1,L-2 regularization, contrary to prevailing literature, indeed exhibits favorable performance in structured pruning tasks. Motivated by this insight, we introduce a hybrid group regularization approach that integrates L-1,L-2 regularization and group L-1/2 regularization (denoted as HGL(1,2) & L-1/2). This novel method addresses the challenge of selecting appropriate L-p regularizers for penalizing weight groups by leveraging L-1,L-2 regularization for penalizing groups with magnitudes exceeding a critical threshold while employing group L-1/2 regularization for other groups. Experimental evaluations are conducted to verify the efficiency of the proposed hybrid group regularization method and the viability of the introduced framework.
引用
收藏
页码:7434 / 7444
页数:11
相关论文
共 54 条
  • [1] Aghasi A., 2017, PROADV NEURAL INF PR, V30
  • [2] Fast Convex Pruning of Deep Neural Networks
    Aghasi, Alireza
    Abdi, Afshin
    Romberg, Justin
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (01): : 158 - 188
  • [3] Alvarez JM, 2016, ADV NEUR IN, V29
  • [4] Structured Pruning of Deep Convolutional Neural Networks
    Anwar, Sajid
    Hwang, Kyuyeon
    Sung, Wonyong
    [J]. ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2017, 13 (03)
  • [5] Blalock D., 2020, Machine Learning and Systems, V2, P129, DOI DOI 10.48550/ARXIV.2003.03033
  • [6] Structured Sparsity of Convolutional Neural Networks via Nonconvex Sparse Group Regularization
    Bui, Kevin
    Park, Fredrick
    Zhang, Shuai
    Qi, Yingyong
    Xin, Jack
    [J]. FRONTIERS IN APPLIED MATHEMATICS AND STATISTICS, 2021, 6
  • [7] Chen Congliang, 2022, J MACH LEARN RES, V23
  • [8] Individual Surveillance Around Parked Aircraft at Nighttime: Thermal Infrared Vision-Based Human Action Recognition
    Ding, Meng
    Ding, Yuanyuan
    Wei, Li
    Xu, Yiming
    Cao, Yunfeng
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (02): : 1084 - 1094
  • [9] France J, 2018, HIST WARFARE, V116, P1, DOI 10.1163/9789004349599_002
  • [10] Glorot X., 2010, Understanding the difficulty of training deep feedforward neural networks