Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning

被引:6
作者
Wang, Bokun [1 ]
Yuan, Zhuoning [2 ]
Ying, Yiming [3 ]
Yang, Tianbao [1 ]
机构
[1] Texas A&M Univ, Dept Comp Sci & Engn, College Stn, TX 77843 USA
[2] Univ Iowa, Dept Comp Sci, Iowa City, IA 52242 USA
[3] SUNY Albany, Dept Math & Stat, Albany, NY 12222 USA
关键词
Meta-Learning; Federated Learning; Model-Agnostic Meta-Learning; Personalized Federated Learning; Memory-Based Algorithms;
D O I
10.48550/arXiv.2106.04911
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, model-agnostic meta-learning (MAML) has become a popular research area. However, the stochastic optimization of MAML is still underdeveloped. Existing MAML algorithms rely on the "episode" idea by sampling a few tasks and data points to update the meta-model at each iteration. Nonetheless, these algorithms either fail to guarantee convergence with a constant mini-batch size or require processing a large number of tasks at every iteration, which is unsuitable for continual learning or cross-device federated learning where only a small number of tasks are available per iteration or per round. To address these issues, this paper proposes memory-based stochastic algorithms for MAML that converge with vanishing error. The proposed algorithms require sampling a constant number of tasks and data samples per iteration, making them suitable for the continual learning scenario. Moreover, we introduce a communication-efficient memory-based MAML algorithm for personalized federated learning in cross-device (with client sampling) and cross-silo (without client sampling) settings. Our theoretical analysis improves the optimization theory for MAML, and our empirical results corroborate our theoretical findings. Interested readers can access our code at https://github.com/bokun-wang/moml.
引用
收藏
页数:46
相关论文
共 53 条
[1]  
Alet F, 2021, ADV NEUR IN
[2]  
Antoniou A, 2019, Arxiv, DOI arXiv:1810.09502
[3]  
Behl H. S., 2019, arXiv
[4]  
Chen Tianyi, 2020, arXiv
[5]  
Chen Tianyi, 2021, arXiv
[6]  
Deng Y., 2020, ADV NEURAL INFORM PR, Vvol33, P15111
[7]  
Dinh CT, 2020, ADV NEUR IN, V33
[8]  
Fallah A, 2020, Arxiv, DOI arXiv:2002.07948
[9]  
Fallah A, 2020, PR MACH LEARN RES, V108, P1082
[10]  
Finn C, 2019, Arxiv, DOI arXiv:1806.02817