Transfer-Learning-Based Gaussian Mixture Model for Distributed Clustering

被引:16
作者
Wang, Rongrong [1 ]
Han, Shiyuan [1 ]
Zhou, Jin [1 ]
Chen, Yuehui [1 ]
Wang, Lin [1 ]
Du, Tao [1 ]
Ji, Ke [1 ]
Zhao, Ya-ou [1 ]
Zhang, Kun [1 ]
机构
[1] Univ Jinan, Shandong Prov Key Lab Network Based Intelligent C, Jinan 250022, Peoples R China
基金
中国国家自然科学基金;
关键词
Clustering algorithms; Distributed databases; Peer-to-peer computing; Transfer learning; Clustering methods; Maximum likelihood estimation; Gaussian mixture model; Distributed clustering; Gaussian mixture model (GMM); peer-to-peer networks; transfer learning; EXPECTATION-MAXIMIZATION; PARAMETER-ESTIMATION; MAXIMUM-LIKELIHOOD; DENSITY-ESTIMATION; EM ALGORITHM;
D O I
10.1109/TCYB.2022.3177242
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distributed clustering based on the Gaussian mixture model (GMM) has exhibited excellent clustering capabilities in peer-to-peer (P2P) networks. However, more iterative numbers and communication overhead are required to achieve the consensus in existing distributed GMM clustering algorithms. In addition, the truth that it cannot find a closed form for the update of parameters in GMM causes the imprecise clustering accuracy. To solve these issues, by utilizing the transfer learning technique, a general transfer distributed GMM clustering framework is exploited to promote the clustering performance and accelerate the clustering convergence. In this work, each node is treated as both the source domain and the target domain, and these nodes can learn from each other to complete the clustering task in distributed P2P networks. Based on this framework, the transfer distributed expectation-maximization algorithm with the fixed learning rate is first presented for data clustering. Then, an improved version is designed to obtain the stable clustering accuracy, in which an adaptive transfer learning strategy is adopted to adjust the learning rate automatically instead of a fixed value. To demonstrate the extensibility of the proposed framework, a representative GMM clustering method, the entropy-type classification maximum-likelihood algorithm, is further extended to the transfer distributed counterpart. Experimental results verify the effectiveness of the presented algorithms in contrast with the existing GMM clustering approaches.
引用
收藏
页码:7058 / 7070
页数:13
相关论文
共 40 条
[1]   Fully Homomorphic based Privacy-Preserving Distributed Expectation Maximization on Cloud [J].
Alabdulatif, Abdulatif ;
Khalil, Ibrahim ;
Zomaya, Albert Y. ;
Tari, Zahir ;
Yi, Xun .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (11) :2668-2681
[2]  
Boyd S.P., 2006, IEEE Transactions on Automatic Control, V51, P1859, DOI DOI 10.1109/TAC.2006.884922
[3]   A method for training finite mixture models under a fuzzy clustering principle [J].
Chatzis, Sotirios .
FUZZY SETS AND SYSTEMS, 2010, 161 (23) :3000-3013
[4]  
Dai W., 2008, Proceedings of the 25th international conference on Machine learning, P200, DOI DOI 10.1145/1390156.1390182
[5]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[6]   Transfer Prototype-Based Fuzzy Clustering [J].
Deng, Zhaohong ;
Jiang, Yizhang ;
Chung, Fu-Lai ;
Ishibuchi, Hisao ;
Choi, Kup-Sze ;
Wang, Shitong .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2016, 24 (05) :1210-1232
[7]  
Dua D., 2019, UCI MACHINE LEARNING
[8]   Ant Colony Stream Clustering: A Fast Density Clustering Algorithm for Dynamic Data Streams [J].
Fahy, Conor ;
Yang, Shengxiang ;
Gongora, Mario .
IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (06) :2215-2228
[9]   Distributed Clustering Using Wireless Sensor Networks [J].
Forero, Pedro A. ;
Cano, Alfonso ;
Giannakis, Georgios B. .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2011, 5 (04) :707-724
[10]   TLPCM: Transfer Learning Possibilistic C-Means [J].
Gargees, Rayan ;
Keller, James M. ;
Popescu, Mihail .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2021, 29 (04) :940-952