Communication-Efficient Federated Linear and Deep Generalized Canonical Correlation Analysis

被引:0
作者
Shrestha, Sagar [1 ]
Fu, Xiao [1 ]
机构
[1] Oregon State Univ, Sch Elect Engn & Comp Sci, Corvallis, OR 97331 USA
基金
美国国家科学基金会;
关键词
Convergence; Signal processing algorithms; Optimization; Federated learning; Correlation; Quantization (signal); Manifolds; Generalized canonical correlation analysis; communication-efficient federated learning; nonconvex optimization; convergence analysis; SUBGRADIENT METHODS; QUANTIZATION; ALGORITHMS; SETS;
D O I
10.1109/TSP.2023.3265886
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Classic and deep learning-based generalized canonical correlation analysis (GCCA) algorithms seek low-dimensional common representations of data entities from multiple "views" (e.g., audio and image) using linear transformations and neural networks, respectively. When the views are acquired and stored at different computing agents (e.g., organizations and edge devices) and data sharing is undesired due to privacy or communication cost considerations, federated learning-based GCCA is well-motivated. In federated learning, the views are kept locally at the agents and only derived, limited information exchange with a central server is allowed. However, applying existing GCCA algorithms onto such setting may incur prohibitively high communication overhead. This work puts forth a communication-efficient federated learning framework for both linear and deep GCCA under the maximum variance (MAX-VAR) formulation. The overhead issue is addressed by aggressively compressing (via quantization) the exchanging information between the computing agents and a central controller. Our synthetic and real-data experiments shows that the proposed algorithm enjoys a substantial reduction of communication overheads with virtually no loss in accuracy and convergence speed compared to the unquantized version. Rigorous convergence analyses are also presented, which is a nontrivial effort since generic federated optimization results do not cover the special problem structure of GCCA. Our result shows that the proposed algorithms for both linear and deep GCCA converge to critical points at a sublinear rate, even under heavy quantization and stochastic approximations. In addition, in the linear MAX-VAR case, the quantized algorithm approaches a global optimum in a geometric rate under reasonable conditions.
引用
收藏
页码:1379 / 1394
页数:16
相关论文
共 67 条
[1]  
AKAHO S, 2001, INT M PSYCH SOC IMPS
[2]  
Alistarh D, 2017, ADV NEUR IN, V30
[3]  
Andrew R., 2013, INT C MACH LEARN, P1247
[4]  
Bach F. R., 2005, 688 U CAL BERK DEP S
[5]  
Basu D, 2019, ADV NEUR IN, V32
[6]  
Benton A, 2019, 4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), P1
[7]  
Berkovsky S, 2007, LECT NOTES ARTIF INT, V4511, P355
[8]  
Bernstein J, 2018, PR MACH LEARN RES, V80
[9]   Distributed Canonical Correlation Analysis in Wireless Sensor Networks With Application to Distributed Blind Source Separation [J].
Bertrand, Alexander ;
Moonen, Marc .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2015, 63 (18) :4800-4813
[10]  
cms, 2013, CMS 2008 2010 DE SYN