Distributed Clustering for Cooperative Multi-Task Learning Networks

被引:4
作者
Li, Jiani [1 ]
Wang, Weihan [2 ]
Abbas, Waseem [3 ]
Koutsoukos, Xenofon [1 ]
机构
[1] Vanderbilt Univ, Dept Comp Sci, Nashville, TN 37235 USA
[2] Stevens Inst Technol, Dept Comp Sci, Hoboken, NJ 07030 USA
[3] Univ Texas Dallas, Dept Syst Engn, Richardson, TX 75080 USA
来源
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING | 2023年 / 10卷 / 06期
关键词
Distributed clustering; distributed cooperative learning; multi-agent networks; multi-task learning; OPTIMIZATION; ADAPTATION; ALGORITHMS;
D O I
10.1109/TNSE.2023.3276854
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Distributed learning enables collaborative training of machine learning models across multiple agents by exchanging model parameters without sharing local data. Each agent generates data from distinct but related distributions, and multi-task learning can be effectively used to model related tasks. This article focuses on clustered multi-task learning, where agents are partitioned into clusters with distinct objectives, and agents in the same cluster share the same objective. The structure of such clusters is unknown apriori. Cooperation with the agents in the same cluster is beneficial and improves the overall learning performance. However, indiscriminate cooperation of agents with different objectives leads to undesired outcomes. Accurately capturing the clustering structure benefits the cooperation and offers many practical benefits; for instance, it helps advertising companies better target their ads. This article proposes an adaptive clustering method that allows distributed agents to learn the most appropriate neighbors to collaborate with and form clusters. We prove the convergence of every agent towards its objective and analyze the network learning performance using the proposed clustering method. Further, we present a method of computing combination weights that approximately optimizes the network's learning performance to determine how one should aggregate the neighbors' model parameters after the clustering step. The theoretical analysis is well-validated by the evaluation results using target localization and digits classification, showing that the proposed clustering method outperforms existing distributed clustering methods as well as the case where agents do not cooperate.
引用
收藏
页码:3933 / 3942
页数:10
相关论文
共 44 条
[1]  
Anguita D., 2013, 21 EUROPEAN S ARTIFI, V3, P437
[2]  
Berman A., 1994, Classics Appl. Math, P9, DOI [10.1137/1.9781611971262, DOI 10.1137/1.9781611971262]
[3]  
Bhaskara A., 2018, P INT C MACH LEARN, P570
[4]   Optimization Methods for Large-Scale Machine Learning [J].
Bottou, Leon ;
Curtis, Frank E. ;
Nocedal, Jorge .
SIAM REVIEW, 2018, 60 (02) :223-311
[5]  
Brendan M.H., 2016, arXiv, DOI [10.48550/arXiv.1602.05629, DOI 10.48550/ARXIV.1602.05629]
[6]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[7]   Diffusion LMS Over Multitask Networks [J].
Chen, Jie ;
Richard, Cedric ;
Sayed, Ali H. .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2015, 63 (11) :2733-2748
[8]   Multitask Diffusion Adaptation Over Networks [J].
Chen, Jie ;
Richard, Cedric ;
Sayed, Ali H. .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2014, 62 (16) :4129-4144
[9]   FedHealth: A Federated Transfer Learning Framework for Wearable Healthcare [J].
Chen, Yiqiang ;
Qin, Xin ;
Wang, Jindong ;
Yu, Chaohui ;
Gao, Wen .
IEEE INTELLIGENT SYSTEMS, 2020, 35 (04) :83-93
[10]   A Survey of Clustering Algorithms for Big Data: Taxonomy and Empirical Analysis [J].
Fahad, Adil ;
Alshatri, Najlaa ;
Tari, Zahir ;
Alamri, Abdullah ;
Khalil, Ibrahim ;
Zomaya, Albert Y. ;
Foufou, Sebti ;
Bouras, Abdelaziz .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2014, 2 (03) :267-279