Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning

被引:0
作者
Zhang, Ce [1 ]
Yao, Xiao [1 ]
Shi, Changfeng [2 ]
Gu, Min [3 ]
机构
[1] Hohai Univ, Coll IoT Engn, Changzhou, Peoples R China
[2] HoHai Univ, Business Sch, ChangZhou, Peoples R China
[3] First Peoples Hosp Changzhou, Changzhou, Peoples R China
关键词
Machine learning; Few-shot learning; K-FAC; Second-order optimization; Adaptive learning rate;
D O I
10.1007/s00530-023-01159-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Model-agnostic meta-learning (MAML) highlights the ability to quickly adapt to new tasks with only a small amount of labeled training data among many few-shot learning algorithms. However, the computational complexity is high, because the MAML algorithm generates a large number of second-order parameters in the secondary gradient update. In addition, due to the non-convex nature of the neural network, the loss landscape has many flat areas, leading to slow convergence during training, and excessively long training. In this paper, a second-order optimization method called Kronecker-factored Approximate Curvature (K-FAC) is proposed to approximate Natural Gradient Descent. K-FAC reduces the computational complexity by approximating the large matrix of the Fisher information as the Kronecker product of two much smaller matrices, and the second-order parameter information is fully utilized to accelerate the convergence. Moreover, in order to solve the problem that Natural Gradient Descent is sensitive to the learning rate, this paper proposes Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning (AK-MAML), which automatically adjusts the learning rate according to the curvature and improves the efficiency of training. Experimental results show that AK-MAML has the ability of faster convergence, lower computation, and higher accuracy on few-shot datasets.
引用
收藏
页码:3169 / 3177
页数:9
相关论文
共 27 条
  • [21] Few-Shot Network Intrusion Detection Based on Model-Agnostic Meta-Learning with L2F Method
    Shi, Zhixin
    Xing, Mengyan
    Zhang, Jing
    Wu, Hao
    2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [22] Question-Answer Methodology for Vulnerable Source Code Review via Prototype-Based Model-Agnostic Meta-Learning
    Corona-Fraga, Pablo
    Hernandez-Suarez, Aldo
    Sanchez-Perez, Gabriel
    Toscano-Medina, Linda Karina
    Perez-Meana, Hector
    Portillo-Portillo, Jose
    Olivares-Mercado, Jesus
    Villalba, Luis Javier Garcia
    FUTURE INTERNET, 2025, 17 (01)
  • [23] Few-shot pump anomaly detection via Diff-WRN-based model-agnostic meta-learning strategy
    Zou, Fengqian
    Sang, Shengtian
    Jiang, Ming
    Li, Xiaoming
    Zhang, Haifeng
    STRUCTURAL HEALTH MONITORING-AN INTERNATIONAL JOURNAL, 2023, 22 (04): : 2674 - 2687
  • [24] Mi-maml: classifying few-shot advanced malware using multi-improved model-agnostic meta-learning
    Ji, Yulong
    Zou, Kunjin
    Zou, Bin
    CYBERSECURITY, 2024, 7 (01):
  • [25] Bayesian optimization-based Model-Agnostic Meta-Learning: Application to predict maximum cyclic moment resistance of steel bolted T-stub connections
    Shen, Yanfei
    Li, Mao
    Li, Yong
    THIN-WALLED STRUCTURES, 2024, 204
  • [26] Adaptive Regularized Warped Gradient Descent Enhances Model Generalization and Meta-learning for Few-shot Learning
    Rao, Shuzhen
    Huang, Jun
    Tang, Zengming
    NEUROCOMPUTING, 2023, 537 : 271 - 281
  • [27] A Task-Adaptive Parameter Transformation Scheme for Model-Agnostic-Meta-Learning-Based Few-Shot Animal Sound Classification
    Moon, Jaeuk
    Kim, Eunbeen
    Hwang, Junha
    Hwang, Eenjun
    APPLIED SCIENCES-BASEL, 2024, 14 (03):