Gray Learning From Non-IID Data With Out-of-Distribution Samples

被引:0
|
作者
Zhao, Zhilin [1 ,2 ]
Cao, Longbing [1 ,2 ]
Wang, Chang-Dong [3 ,4 ]
机构
[1] Macquarie Univ, Data Sci Lab, Sch Comp, Sydney, NSW 2109, Australia
[2] Macquarie Univ, Data Sci Lab, DataX Res Ctr, Sydney, NSW 2109, Australia
[3] Sun Yat Sen Univ, Comp Sci & Engn, Guangdong Prov Key Lab Computat Sci, Minist Educ, Guangzhou 510275, Peoples R China
[4] Sun Yat Sen Univ, Comp Sci & Engn, Key Lab Machine Intelligence & Adv Comp, Minist Educ, Guangzhou 510275, Peoples R China
基金
澳大利亚研究理事会;
关键词
Training; Noise measurement; Task analysis; Complexity theory; Training data; Neural networks; Metalearning; Complementary label; generalization; gray learning (GL); non-independent and identically distributed (Non-IID) data; out-of-distribution data;
D O I
10.1109/TNNLS.2023.3330475
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The integrity of training data, even when annotated by experts, is far from guaranteed, especially for non-independent and identically distributed (non-IID) datasets comprising both in- and out-of-distribution samples. In an ideal scenario, the majority of samples would be in-distribution, while samples that deviate semantically would be identified as out-of-distribution and excluded during the annotation process. However, experts may erroneously classify these out-of-distribution samples as in-distribution, assigning them labels that are inherently unreliable. This mixture of unreliable labels and varied data types makes the task of learning robust neural networks notably challenging. We observe that both in- and out-of-distribution samples can almost invariably be ruled out from belonging to certain classes, aside from those corresponding to unreliable ground-truth labels. This opens the possibility of utilizing reliable complementary labels that indicate the classes to which a sample does not belong. Guided by this insight, we introduce a novel approach, termed gray learning (GL), which leverages both ground-truth and complementary labels. Crucially, GL adaptively adjusts the loss weights for these two label types based on prediction confidence levels. By grounding our approach in statistical learning theory, we derive bounds for the generalization error, demonstrating that GL achieves tight constraints even in non-IID settings. Extensive experimental evaluations reveal that our method significantly outperforms alternative approaches grounded in robust statistics.
引用
收藏
页码:1396 / 1409
页数:14
相关论文
共 50 条
  • [21] IOFL: Intelligent-Optimization-Based Federated Learning for Non-IID Data
    Li, Xinyan
    Zhao, Huimin
    Deng, Wu
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (09): : 16693 - 16699
  • [22] Non-IID Federated Learning With Sharper Risk Bound
    Wei, Bojian
    Li, Jian
    Liu, Yong
    Wang, Weiping
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (05) : 6906 - 6917
  • [23] Detecting Out-of-Distribution Data in Wireless Communications Applications of Deep Learning
    Liu, Jinshan
    Oyedare, Taiwo
    Park, Jung-Min
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (04) : 2476 - 2487
  • [24] Improving Out-of-Distribution Detection by Learning From the Deployment Environment
    Inkawhich, Nathan
    Zhang, Jingyang
    Davis, Eric K.
    Luley, Ryan
    Chen, Yiran
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 : 2070 - 2086
  • [25] Long-Term Client Selection for Federated Learning With Non-IID Data: A Truthful Auction Approach
    Tan, Jinghong
    Liu, Zhian
    Guo, Kun
    Zhao, Mingxiong
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (05): : 4953 - 4970
  • [26] Coalitional Federated Learning: Improving Communication and Training on Non-IID Data With Selfish Clients
    Arisdakessian, Sarhad
    Wahab, Omar Abdel
    Mourad, Azzam
    Otrok, Hadi
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2023, 16 (04) : 2462 - 2476
  • [27] FedSea: Federated Learning via Selective Feature Alignment for Non-IID Multimodal Data
    Tan, Min
    Feng, Yinfu
    Chu, Lingqiang
    Shi, Jingcheng
    Xiao, Rong
    Tang, Haihong
    Yu, Jun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 5807 - 5822
  • [28] Digital Twin-Empowered Federated Incremental Learning for Non-IID Privacy Data
    Wang, Qian
    Chen, Siguang
    Wu, Meng
    Li, Xue
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2025, 24 (05) : 3860 - 3877
  • [29] Federated Multi-Task Learning on Non-IID Data Silos: An Experimental Study
    Yang, Yuwen
    Lu, Yuxiang
    Huang, Suizhi
    Sirejiding, Shalayiding
    Lu, Hongtao
    Ding, Yue
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 684 - 693
  • [30] Diffusion Policies for Out-of-Distribution Generalization in Offline Reinforcement Learning
    Ada, Suzan Ece
    Oztop, Erhan
    Ugur, Emre
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (04) : 3116 - 3123