Meta-Learning with a Geometry-Adaptive Preconditioner

被引:14
|
作者
Kang, Suhyun [1 ]
Hwang, Duhun [1 ]
Eo, Moonjung [1 ]
Kim, Taesup [2 ]
Rhee, Wonjong [1 ,3 ,4 ]
机构
[1] Seoul Natl Univ, Dept Intelligence & Informat, Seoul, South Korea
[2] Seoul Natl Univ, Grad Sch Data Sci, Seoul, South Korea
[3] Seoul Natl Univ, IPAI, Seoul, South Korea
[4] Seoul Natl Univ, AIIS, Seoul, South Korea
关键词
D O I
10.1109/CVPR52729.2023.01543
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model-agnostic meta-learning (MAML) is one of the most successful meta-learning algorithms. It has a bi-level optimization structure where the outer-loop process learns a shared initialization and the inner-loop process optimizes task-specific weights. Although MAML relies on the standard gradient descent in the inner-loop, recent studies have shown that controlling the inner-loop's gradient descent with a meta-learned preconditioner can be beneficial. Existing preconditioners, however, cannot simultaneously adapt in a task-specific and path-dependent way. Additionally, they do not satisfy the Riemannian metric condition, which can enable the steepest descent learning with preconditioned gradient. In this study, we propose Geometry-Adaptive Pre-conditioned gradient descent (GAP) that can overcome the limitations in MAML; GAP can efficiently meta-learn a preconditioner that is dependent on task-specific parameters, and its preconditioner can be shown to be a Riemannian metric. Thanks to the two properties, the geometryadaptive preconditioner is effective for improving the innerloop optimization. Experiment results show that GAP outperforms the state-of-the-art MAML family and preconditioned gradient descent-MAML (PGD-MAML) family in a variety of few-shot learning tasks. Code is available at: https://github.com/ Suhyun777/CVPR23-GAP.
引用
收藏
页码:16080 / 16090
页数:11
相关论文
共 50 条
  • [31] Bypassing the scaling relations in oxygen electrocatalysis with geometry-adaptive catalysts
    Cepitis, Ritums
    Ivanistsev, Vladislav
    Rossmeisl, Jan
    Kongi, Nadezda
    CATALYSIS SCIENCE & TECHNOLOGY, 2024, 14 (08) : 2105 - 2113
  • [32] Adaptive-Control-Oriented Meta-Learning for Nonlinear Systems
    Richards, Spencer M.
    Azizan, Navid
    Slotine, Jean-Jacques
    Pavone, Marco
    ROBOTICS: SCIENCE AND SYSTEM XVII, 2021,
  • [33] Meta-learning in Reinforcement Learning
    Schweighofer, N
    Doya, K
    NEURAL NETWORKS, 2003, 16 (01) : 5 - 9
  • [34] Learning to Forget for Meta-Learning
    Baik, Sungyong
    Hong, Seokil
    Lee, Kyoung Mu
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 2376 - 2384
  • [35] Learning to Learn Better Unimodal Representations via Adaptive Multimodal Meta-Learning
    Sun, Ya
    Mai, Sijie
    Hu, Haifeng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2209 - 2223
  • [36] Learning With Dual Demonstration Domains: Random Domain-Adaptive Meta-Learning
    Hu, Ziye
    Li, Wei
    Gan, Zhongxue
    Guo, Weikun
    Zhu, Jiwei
    Gao, Xiang
    Yang, Xuyun
    Peng, Yueyan
    Zuo, Zhihao
    Wen, James Zhiqing
    Zhou, Decheng
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3523 - 3530
  • [37] Submodular Meta-Learning
    Adibi, Arman
    Mokhtari, Aryan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [38] Online Meta-Learning
    Finn, Chelsea
    Rajeswaran, Aravind
    Kakade, Sham
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [39] Geometry-Adaptive Network for Robust Detection of Placenta Accreta Spectrum Disorders
    Chen, Zailiang
    Zhu, Jiang
    Shen, Hailan
    Liu, Hui
    Li, Yajing
    Zhao, Rongchang
    Yu, Feiyang
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT VII, 2023, 14226 : 43 - 53
  • [40] Meta-Learning With Adaptive Learning Rates for Few-Shot Fault Diagnosis
    Chang, Liang
    Lin, Yan-Hui
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2022, 27 (06) : 5948 - 5958