Slimmable Dataset Condensation

被引:37
作者
Liu, Songhua [1 ]
Ye, Jingwen [1 ]
Yu, Runpeng [1 ]
Wang, Xinchao [1 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR | 2023年
基金
新加坡国家研究基金会;
关键词
CHALLENGES;
D O I
10.1109/CVPR52729.2023.00366
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dataset distillation, also known as dataset condensation, aims to compress a large dataset into a compact synthetic one. Existing methods perform dataset condensation by assuming a fixed storage or transmission budget. When the budget changes, however, they have to repeat the synthesizing process with access to original datasets, which is highly cumbersome if not infeasible at all. In this paper, we explore the problem of slimmable dataset condensation, to extract a smaller synthetic dataset given only previous condensation results. We first study the limitations of existing dataset condensation algorithms on such a successive compression setting and identify two key factors: (1) the inconsistency of neural networks over different compression times and (2) the underdetermined solution space for synthetic data. Accordingly, we propose a novel training objective for slimmable dataset condensation to explicitly account for both factors. Moreover, synthetic datasets in our method adopt a significance-aware parameterization. Theoretical derivation indicates that an upper-bounded error can be achieved by discarding the minor components without training. Alternatively, if training is allowed, this strategy can serve as a strong initialization that enables a fast convergence. Extensive comparisons and ablations demonstrate the superiority of the proposed solution over existing methods on multiple benchmarks.
引用
收藏
页码:3759 / 3768
页数:10
相关论文
共 59 条
[21]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[22]  
Lee Hae Beom, 2022, ARXIV220800719
[23]  
Lee Saehyung, 2022, ARXIV220202916
[24]   Federated Learning: Challenges, Methods, and Future Directions [J].
Li, Tian ;
Sahu, Anit Kumar ;
Talwalkar, Ameet ;
Smith, Virginia .
IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (03) :50-60
[25]  
Liu Ping, 2022, ARXIV220914851
[26]  
Liu S, 2022, P ADV NEUR INF PROC
[27]  
Lorraine J, 2020, PR MACH LEARN RES, V108, P1540
[28]  
Maclaurin D, 2015, PR MACH LEARN RES, V37, P2113
[29]   Online continual learning in image classification: An empirical survey [J].
Mai, Zheda ;
Li, Ruiwen ;
Jeong, Jihwan ;
Quispe, David ;
Kim, Hyunwoo ;
Sanner, Scott .
NEUROCOMPUTING, 2022, 469 :28-51
[30]  
Masarczyk Wojciech, 2020, P IEEE CVF C COMP VI