Remember the Past: Distilling Datasets into Addressable Memories for Neural Networks

被引:0
|
作者
Deng, Zhiwei [1 ]
Russakovsky, Olga [1 ]
机构
[1] Princeton Univ, Dept Comp Sci, Princeton, NJ 08544 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose an algorithm that compresses the critical information of a large dataset into compact addressable memories. These memories can then be recalled to quickly re-train a neural network and recover the performance (instead of storing and re-training on the full original dataset). Building upon the dataset distillation framework, we make a key observation that a shared common representation allows for more efficient and effective distillation. Concretely, we learn a set of bases (aka "memories") which are shared between classes and combined through learned flexible addressing functions to generate a diverse set of training examples. This leads to several benefits: 1) the size of compressed data does not necessarily grow linearly with the number of classes; 2) an overall higher compression rate with more effective distillation is achieved; and 3) more generalized queries are allowed beyond recalling the original classes. We demonstrate state-of-the-art results on the dataset distillation task across six benchmarks, including up to 16.5% and 9.7% in retained accuracy improvement when distilling CIFAR10 and CIFAR100 respectively. We then leverage our framework to perform continual learning, achieving state-of-the-art results on four benchmarks, with 23.2% accuracy improvement on MANY. The code is released on our project webpage(1).
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Dynamics of neural networks with delay: Attractors and content-addressable memory
    Wu, JH
    DIFFERENTIAL EQUATIONS AND NONLINEAR MECHANICS, 2001, 528 : 401 - 417
  • [32] Distilling Knowledge From an Ensemble of Convolutional Neural Networks for Seismic Fault Detection
    Wang, Zirui
    Li, Bo
    Liu, Naihao
    Wu, Bangyu
    Zhu, Xu
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [33] Neural Networks and Multimedia Datasets: Estimating the Size of Neural Networks for Achieving High Classification Accuracy
    Lappas, Georgios
    MUSP '06: PROCEEDINGS OF THE 9TH WSEAS INTERNATIONAL CONFERENCE ON MULTIMEDIA SYSTEMS AND SIGNAL PROCESSING, 2009, : 237 - +
  • [34] Predicting Diabetes in Imbalanced Datasets using Neural Networks
    Guan, Hannah
    Zhang, Chonghao
    13TH ACM INTERNATIONAL CONFERENCE ON BIOINFORMATICS, COMPUTATIONAL BIOLOGY AND HEALTH INFORMATICS, BCB 2022, 2022,
  • [35] Twin Neural Networks for the classification of large unbalanced datasets
    Jayadeva
    Pant, Himanshu
    Sharma, Mayank
    Soman, Sumit
    NEUROCOMPUTING, 2019, 343 : 34 - 49
  • [36] Datasets and Interfaces for Benchmarking Heterogeneous Graph Neural Networks
    Liu, Yijian
    Zhang, Hongyi
    Yang, Cheng
    Li, Ao
    Ji, Yugang
    Zhang, Luhao
    Li, Tao
    Yang, Jinyu
    Zhao, Tianyu
    Yang, Juan
    Huang, Hai
    Shi, Chuan
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 5346 - 5350
  • [37] Discovering trends in large datasets using neural networks
    Kaikhah, K
    Doddameti, S
    APPLIED INTELLIGENCE, 2006, 24 (01) : 51 - 60
  • [38] Artificial Neural Networks with Random Weights for Incomplete Datasets
    Mesquita, Diego P. P.
    Gomes, Joao Paulo P.
    Rodrigues, Leonardo R.
    NEURAL PROCESSING LETTERS, 2019, 50 (03) : 2345 - 2372
  • [39] Fluorescence microscopy datasets for training deep neural networks
    Hagen, Guy M.
    Bendesky, Justin
    Machado, Rosa
    Tram-Anh Nguyen
    Kumar, Tanmay
    Ventura, Jonathan
    GIGASCIENCE, 2021, 10 (05):
  • [40] Artificial Neural Networks with Random Weights for Incomplete Datasets
    Diego P. P. Mesquita
    João Paulo P. Gomes
    Leonardo R. Rodrigues
    Neural Processing Letters, 2019, 50 : 2345 - 2372