Efficient Variance Reduction for Meta-Learning

被引:0
|
作者
Yang, Hansi [1 ]
Kwok, James T. [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Clear Water Bay, Hong Kong, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning tries to learn meta-knowledge from a large number of tasks. However, the stochastic meta-gradient can have large variance due to data sampling (from each task) and task sampling (from the whole task distribution), leading to slow convergence. In this paper, we propose a novel approach that integrates variance reduction with first-order meta-learning algorithms such as Reptile. It retains the bilevel formulation which better captures the structure of meta-learning, but does not require storing the vast number of task-specific parameters in general bilevel variance reduction methods. Theoretical results show that it has fast convergence rate due to variance reduction. Experiments on benchmark few-shot classification data sets demonstrate its effectiveness over state-of-the-art meta-learning algorithms with and without variance reduction.
引用
收藏
页数:26
相关论文
共 50 条
  • [1] Meta-learning Control Variates: Variance Reduction with Limited Data
    Sun, Zhuo
    Oates, Chris J.
    Briol, Francois-Xavier
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 2047 - 2057
  • [2] TinyMetaFed: Efficient Federated Meta-learning for TinyML
    Ren, Haoyu
    Li, Xue
    Anicic, Darko
    Runkler, Thomas A.
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2023, PT IV, 2025, 2136 : 126 - 137
  • [3] Meta-learning for efficient unsupervised domain adaptation
    Vettoruzzo, Anna
    Bouguelia, Mohamed-Rafik
    Roegnvaldsson, Thorsteinn
    NEUROCOMPUTING, 2024, 574
  • [4] Memory Efficient Meta-Learning with Large Images
    Bronskill, John
    Massiceti, Daniela
    Patacchiola, Massimiliano
    Hofmann, Katja
    Nowozin, Sebastian
    Turner, Richard E.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Efficient Meta-Learning for Continual Learning with Taylor Expansion Approximation
    Zou, Xiaohan
    Lin, Tong
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [6] Towards Sample-efficient Overparameterized Meta-learning
    Sun, Yue
    Narang, Adhyyan
    Gulluk, Halil Ibrahim
    Oymak, Samet
    Fazel, Maryam
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [7] AeroINR: Meta-learning for Efficient Generation of Aerodynamic Geometries
    Bamford, Tom
    Toal, David
    Keane, Andy
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES-APPLIED DATA SCIENCE TRACK, PT IX, ECML PKDD 2024, 2024, 14949 : 452 - 467
  • [8] Learning from Past Observations: Meta-Learning for Efficient Clustering Analyses
    Fritz, Manuel.
    Tschechlov, Dennis
    Schwarz, Holger
    BIG DATA ANALYTICS AND KNOWLEDGE DISCOVERY (DAWAK 2020), 2020, 12393 : 364 - 379
  • [9] Learning Meta-Learning (LML) dataset: Survey data of meta-learning parameters
    Corraya, Sonia
    Al Mamun, Shamim
    Kaiser, M. Shamim
    DATA IN BRIEF, 2023, 51
  • [10] Meta-learning in Reinforcement Learning
    Schweighofer, N
    Doya, K
    NEURAL NETWORKS, 2003, 16 (01) : 5 - 9