Up to Thousands-fold Storage Saving: Towards Efficient Data-Free Distillation of Large-Scale Visual Classifiers

被引:0
|
作者
Ye, Fanfan [1 ,2 ]
Lu, Bingyi [3 ]
Ma, Liang [2 ]
Zhong, Qiaoyong [2 ]
Xie, Di [2 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Hikvis Res Inst, Hangzhou, Peoples R China
[3] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023 | 2023年
关键词
Data-free; Knowledge Distillation; Storage Efficient;
D O I
10.1145/3581783.3611748
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data-Free Knowledge Distillation (DFKD) has started to make breakthroughs in classification tasks for large-scale datasets such as ImageNet-1k. Despite the encouraging results achieved, these modern DFKD methods still suffer from the massive waste of system storage and I/O resources. They either synthesize and store a vast amount of pseudo data or build thousands of generators. In this work, we introduce a storage-efficient scheme called Class-Expanding DFKD (CE-DFKD). It allows us to reduce storage costs by orders of magnitude in large-scale tasks using just one or a few generators without explicitly storing any data. The key to the success of our approach lies in alleviating the mode collapse of the generator by expanding its collapse range. Specifically, we first investigate and address the optimization conflict of previous single-generator-based DFKD methods by introducing conditional constraints. Then, we propose two class-expanding strategies to enrich the conditional information of the generator from both inter-class and intra-class perspectives. With the diversity of generated samples significantly enhanced, the proposed CE-DFKD outperforms existing methods by a large margin while achieving up to thousands of times storage savings. Besides the ImageNet-1k, the proposed CE-DFKD is compatible with widely used small-scale datasets and can be scaled to the more complex ImageNet-21k-P dataset, which was previously unreported in prior DFKD methods.
引用
收藏
页码:8376 / 8386
页数:11
相关论文
共 1 条
  • [1] A mechanics-based data-free Problem Independent Machine Learning (PIML) model for large-scale structural analysis and design optimization
    Huang, Mengcheng
    Liu, Chang
    Guo, Yilin
    Zhang, Linfeng
    Du, Zongliang
    Guo, Xu
    JOURNAL OF THE MECHANICS AND PHYSICS OF SOLIDS, 2024, 193