Unsupervised Federated Optimization at the Edge: D2D-Enabled Learning Without Labels

被引:1
作者
Wagle, Satyavrat [1 ]
Hosseinalipour, Seyyedali [2 ]
Khosravan, Naji [3 ]
Brinton, Christopher G. [1 ]
机构
[1] Purdue Univ, Sch Elect & Comp Engn, W Lafayette, IN 47907 USA
[2] Univ Buffalo, SUNY, Elect Engn Dept, Buffalo, NY 14260 USA
[3] Zillow Grp, Seattle, WA 98101 USA
关键词
Information exchange; Data models; Training; Servers; Self-supervised learning; Data privacy; Computational modeling; Federated learning; unsupervised learning; INTERNET; SCHEME;
D O I
10.1109/TCCN.2024.3392792
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Federated learning (FL) is a popular solution for distributed machine learning (ML). While FL has traditionally been studied for supervised ML tasks, in many applications, it is impractical to assume availability of labeled data across devices. To this end, we develop Cooperative Federated unsupervised Contrastive Learning (CF-CL) to facilitate FL across edge devices with unlabeled datasets. CF-CL employs local device cooperation where either explicit (i.e., raw data) or implicit (i.e., embeddings) information is exchanged through device-to-device (D2D) communications to improve local diversity. Specifically, we introduce a smart information push-pull methodology for data/embedding exchange tailored to FL settings with either soft or strict data privacy restrictions. Information sharing is conducted through a probabilistic importance sampling technique at receivers leveraging a carefully crafted reserve dataset provided by transmitters. In the implicit case, embedding exchange is further integrated into the local ML training at the devices via a regularization term incorporated into the contrastive loss, augmented with a dynamic contrastive margin to adjust the volume of latent space explored. Numerical evaluations demonstrate that CF-CL leads to alignment of latent spaces learned across devices, results in faster and more efficient global model training, and is effective in extreme non-i.i.d. data distribution settings across devices.
引用
收藏
页码:2252 / 2268
页数:17
相关论文
共 53 条
  • [1] MHADBOR: AI-Enabled Administrative-Distance-Based Opportunistic Load Balancing Scheme for an Agriculture Internet of Things Network
    Adil, Muhammad
    Khan, Muhammad Khurram
    Jamjoom, Mona
    Farouk, Ahmed
    [J]. IEEE MICRO, 2022, 42 (01) : 41 - 50
  • [2] Federated Learning for Privacy Preservation in Smart Healthcare Systems: A Comprehensive Survey
    Ali, Mansoor
    Naeem, Faisal
    Tariq, Muhammad
    Kaddoum, Georges
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2023, 27 (02) : 778 - 789
  • [3] Arthur D, 2007, PROCEEDINGS OF THE EIGHTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, P1027
  • [4] Azam SS, 2022, PR MACH LEARN RES, V151
  • [5] Bercea CI, 2021, Arxiv, DOI arXiv:2103.03705
  • [6] Bistritz I., 2020, P 34 INT C NEUR INF, P22593
  • [7] Federated learning with hierarchical clustering of local updates to improve training on non-IID data
    Briggs, Christopher
    Fan, Zhong
    Andras, Peter
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [8] Bui D, 2019, Arxiv, DOI arXiv:1909.12535
  • [9] Relay-Assisted Federated Edge Learning: Performance Analysis and System Optimization
    Chen, Lunyuan
    Fan, Lisheng
    Lei, Xianfu
    Duong, Trung Q.
    Nallanathan, Arumugam
    Karagiannidis, George K.
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2023, 71 (06) : 3387 - 3401
  • [10] Chen T, 2020, PR MACH LEARN RES, V119