Reusing Convolutional Neural Network Models through Modularization and Composition

被引:2
|
作者
Qi, Binhang [1 ]
Sun, Hailong [2 ]
Zhang, Hongyu [3 ]
Gao, Xiang [4 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, SKLSDE, Xueyuan Rd,Haidian Dist 37, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Software, SKLSDE, Xueyuan Rd,Haidian Dist 37, Beijing 100191, Peoples R China
[3] Chongqing Univ, Sch Big Data Software Engn, 55 Univ Town South Rd, Chongqing 401331, Peoples R China
[4] Beihang Univ, Sch Software, Xueyuan Rd,Haidian Dist 37, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Model reuse; convolutional neural network; CNN modularization; module composition;
D O I
10.1145/3632744
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
With the widespread success of deep learning technologies, many trained deep neural network (DNN) models are now publicly available. However, directly reusing the public DNN models for new tasks often fails due to mismatching functionality or performance. Inspired by the notion of modularization and composition in software reuse, we investigate the possibility of improving the reusability of DNN models in a more fine-grained manner. Specifically, we propose two modularization approaches named CNNSplitter and GradSplitter, which can decompose a trained convolutional neural network (CNN) model for N-class classification into N small reusable modules. Each module recognizes one of the N classes and contains a part of the convolution kernels of the trained CNN model. Then, the resulting modules can be reused to patch existing CNN models or build new CNN models through composition. The main difference between CNNSplitter and GradSplitter lies in their search methods: the former relies on a genetic algorithm to explore search space, while the latter utilizes a gradient-based search method. Our experiments with three representative CNNs on three widely used public datasets demonstrate the effectiveness of the proposed approaches. Compared with CNNSplitter, GradSplitter incurs less accuracy loss, produces much smaller modules (19.88% fewer kernels), and achieves better results on patching weak models. In particular, experiments on GradSplitter show that (1) by patching weak models, the average improvement in terms of precision, recall, and F1-score is 17.13%, 4.95%, and 11.47%, respectively, and (2) for a new task, compared with the models trained from scratch, reusing modules achieves similar accuracy (the average loss of accuracy is only 2.46%) without a costly training process. Our approaches provide a viable solution to the rapid development and improvement of CNN models.
引用
收藏
页数:39
相关论文
共 50 条
  • [1] Reusing Deep Neural Network Models through Model Re-engineering
    Qi, Binhang
    Sun, Hailong
    Gao, Xiang
    Zhang, Hongyu
    Li, Zhaotian
    Liu, Xudong
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE, 2023, : 983 - 994
  • [2] Performance evaluation of various deep convolutional neural network models through classification of malware
    Tasneem, Zareen
    Afnan, Maria
    Hossain, Md. Arman
    Rahman, Md. Mahbubur
    Dey, Samrat Kumar
    INTERNATIONAL JOURNAL OF INFORMATION AND COMPUTER SECURITY, 2023, 21 (3-4) : 414 - 435
  • [3] A Convolutional Neural Network Approach to the Classification of Engineering Models
    Manda, Bharadwaj
    Bhaskare, Pranjal
    Muthuganapathy, Ramanathan
    IEEE ACCESS, 2021, 9 : 22711 - 22723
  • [4] Welding defects classification through a Convolutional Neural Network
    Perri, Stefania
    Spagnolo, Fanny
    Frustaci, Fabio
    Corsonello, Pasquale
    MANUFACTURING LETTERS, 2023, 35 : 29 - 32
  • [5] Improved Breast Cancer Classification Through Combining Graph Convolutional Network and Convolutional Neural Network
    Zhang, Yu-Dong
    Satapathy, Suresh Chandra
    Guttery, David S.
    Manuel Gorriz, Juan
    Wang, Shui-Hua
    INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (02)
  • [6] Are Image Patches Beneficial for Initializing Convolutional Neural Network Models?
    Lehmann, Daniel
    Ebner, Marc
    VISAPP: PROCEEDINGS OF THE 16TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS - VOL. 5: VISAPP, 2021, : 346 - 353
  • [7] Morpheme Segmentation for Russian: Evaluation of Convolutional Neural Network Models
    Maltina, Lyudmila
    Malafeev, Alexey
    ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS (AIST 2019), 2020, 1086 : 160 - 166
  • [8] Comparison of Convolutional Neural Network Models for Document Image Classification
    Yaman, Dogucan
    Eyiokur, Fevziye Irem
    Ekenel, Hazim Kemal
    2017 25TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2017,
  • [9] Optimized Convolutional Neural Network Models for Skin Lesion Classification
    Villa-Pulgarin, Juan Pablo
    Ruales-Torres, Anderson Alberto
    Arias-Garzon, Daniel
    Bravo-Ortiz, Mario Alejandro
    Arteaga-Arteaga, Harold Brayan
    Mora-Rubio, Alejandro
    Alzate-Grisales, Jesus Alejandro
    Mercado-Ruiz, Esteban
    Hassaballah, M.
    Orozco-Arias, Simon
    Cardona-Morales, Oscar
    Tabares-Soto, Reinel
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 70 (02): : 2131 - 2148
  • [10] Comparison of Convolutional Neural Network Models for Food Image Classification
    Ozsert Yigit, Gozde
    Ozyildirim, Buse Melis
    2017 IEEE INTERNATIONAL CONFERENCE ON INNOVATIONS IN INTELLIGENT SYSTEMS AND APPLICATIONS (INISTA), 2017, : 349 - 353