Slimmable Dataset Condensation

被引:37
作者
Liu, Songhua [1 ]
Ye, Jingwen [1 ]
Yu, Runpeng [1 ]
Wang, Xinchao [1 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR | 2023年
基金
新加坡国家研究基金会;
关键词
CHALLENGES;
D O I
10.1109/CVPR52729.2023.00366
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dataset distillation, also known as dataset condensation, aims to compress a large dataset into a compact synthetic one. Existing methods perform dataset condensation by assuming a fixed storage or transmission budget. When the budget changes, however, they have to repeat the synthesizing process with access to original datasets, which is highly cumbersome if not infeasible at all. In this paper, we explore the problem of slimmable dataset condensation, to extract a smaller synthetic dataset given only previous condensation results. We first study the limitations of existing dataset condensation algorithms on such a successive compression setting and identify two key factors: (1) the inconsistency of neural networks over different compression times and (2) the underdetermined solution space for synthetic data. Accordingly, we propose a novel training objective for slimmable dataset condensation to explicitly account for both factors. Moreover, synthetic datasets in our method adopt a significance-aware parameterization. Theoretical derivation indicates that an upper-bounded error can be achieved by discarding the minor components without training. Alternatively, if training is allowed, this strategy can serve as a strong initialization that enables a fast convergence. Extensive comparisons and ablations demonstrate the superiority of the proposed solution over existing methods on multiple benchmarks.
引用
收藏
页码:3759 / 3768
页数:10
相关论文
共 59 条
[1]  
Bradbury J., 2018, JAX COMPOSABLE TRANS
[2]  
Cazenavette George, 2022, ARXIV220311932
[3]  
Chen Can, 2022, 36 C NEUR INF PROC S
[4]  
De Lange Matthias, 2021, Transactions on Pattern Analysis and Machine Intelligence (T-PAMI), V44, P2
[5]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[6]  
Deng Zhiwei, 2022, ARXIV220602916
[7]  
Dong Tian, 2022, ARXIV220600240
[8]  
Dosovitskiy A., 2020, ICLR 2021
[9]  
Elsken Thomas, 2018, Journal of Machine Learning Research
[10]  
Goetz Jack, 2020, ARXIV200804489