Efficient Variance Reduction for Meta-Learning

被引:0
|
作者
Yang, Hansi [1 ]
Kwok, James T. [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Clear Water Bay, Hong Kong, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning tries to learn meta-knowledge from a large number of tasks. However, the stochastic meta-gradient can have large variance due to data sampling (from each task) and task sampling (from the whole task distribution), leading to slow convergence. In this paper, we propose a novel approach that integrates variance reduction with first-order meta-learning algorithms such as Reptile. It retains the bilevel formulation which better captures the structure of meta-learning, but does not require storing the vast number of task-specific parameters in general bilevel variance reduction methods. Theoretical results show that it has fast convergence rate due to variance reduction. Experiments on benchmark few-shot classification data sets demonstrate its effectiveness over state-of-the-art meta-learning algorithms with and without variance reduction.
引用
收藏
页数:26
相关论文
共 50 条
  • [31] Meta-Learning Representations for Continual Learning
    Javed, Khurram
    White, Martha
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [32] TOOLS AND TASKS FOR LEARNING AND META-LEARNING
    Jaworski, Barbara
    JOURNAL OF MATHEMATICS TEACHER EDUCATION, 2005, 8 (05) : 359 - 361
  • [33] Tools and Tasks for Learning and Meta-learning
    Barbara Jaworski
    Journal of Mathematics Teacher Education, 2005, 8 (5) : 359 - 361
  • [34] Meta-Learning the Difference: Preparing Large Language Models for Efficient Adaptation
    Hou, Zejiang
    Salazar, Julian
    Polovets, George
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1249 - 1265
  • [35] EvoGrad: Efficient Gradient-Based Meta-Learning and Hyperparameter Optimization
    Bohdal, Ondrej
    Yang, Yongxin
    Hospedales, Timothy
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [36] Efficient Meta-Learning through Task-Specific Pseudo Labelling
    Lee, Sanghyuk
    Lee, Seunghyun
    Song, Byung Cheol
    ELECTRONICS, 2023, 12 (13)
  • [37] SAMPLE EFFICIENT SUBSPACE-BASED REPRESENTATIONS FOR NONLINEAR META-LEARNING
    Gulluk, Halil Ibrahim
    Sun, Yue
    Oymak, Samet
    Fazel, Maryam
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3685 - 3689
  • [38] On meta-learning rule learning heuristics
    Janssen, Frederik
    Fuernkranz, Johannes
    ICDM 2007: PROCEEDINGS OF THE SEVENTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, 2007, : 529 - 534
  • [39] Learning to Diagnose: Meta-Learning for Efficient Adaptation in Few-Shot AIOps Scenarios
    Duan, Yunfeng
    Bao, Haotong
    Bai, Guotao
    Wei, Yadong
    Xue, Kaiwen
    You, Zhangzheng
    Zhang, Yuantian
    Liu, Bin
    Chen, Jiaxing
    Wang, Shenhuan
    Ou, Zhonghong
    ELECTRONICS, 2024, 13 (11)
  • [40] Probabilistic Active Meta-Learning
    Kaddour, Jean
    Saemundsson, Steindor
    Deisenroth, Marc Peter
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33