Individually Conditional Individual Mutual Information Bound on Generalization Error

被引:8
作者
Zhou, Ruida [1 ]
Tian, Chao [1 ]
Liu, Tie [1 ]
机构
[1] Texas A&M Univ, Dept Elect & Comp Engn, College Stn, TX 77843 USA
基金
美国国家科学基金会;
关键词
Mutual information; Training; Random variables; Heuristic algorithms; Training data; Noise measurement; Upper bound; Information-theoretic bounds; generalization error; stochastic gradient Langevin dynamics;
D O I
10.1109/TIT.2022.3144615
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose an information-theoretic bound on the generalization error based on a combination of the error decomposition technique of Bu et al. and the conditional mutual information (CMI) construction of Steinke and Zakynthinou. In a previous work, Haghifam et al. proposed a different bound combining the two aforementioned techniques, which we refer to as the conditional individual mutual information (CIMI) bound. However, in a simple Gaussian setting, both the CMI and the CIMI bounds are order-wise worse than that by Bu et al. This observation motivated us to propose the bound, which overcomes this issue by reducing the conditioning terms in the conditional mutual information. In the process of establishing this bound, a conditional decoupling lemma is established, which also leads to a meaningful dichotomy and comparison among these information-theoretic bounds. As an application of the proposed bound, we analyze the noisy and iterative stochastic gradient Langevin dynamics and provide an upper bound on its generalization error.
引用
收藏
页码:3304 / 3316
页数:13
相关论文
共 50 条
  • [1] Tightening mutual information-based bounds on generalization error
    Bu Y.
    Zou S.
    Veeravalli V.V.
    IEEE Journal on Selected Areas in Information Theory, 2020, 1 (01): : 121 - 130
  • [2] A Tight Upper Bound on Mutual Information
    Hledik, Michal
    Sokolowski, Thomas R.
    Tkacik, Gasper
    2019 IEEE INFORMATION THEORY WORKSHOP (ITW), 2019, : 70 - 74
  • [3] Error Exponents and α-Mutual Information
    Verdu, Sergio
    ENTROPY, 2021, 23 (02) : 1 - 52
  • [4] MPEG-7 descriptor selection using Localized Generalization Error Model with Mutual Information
    Wang, Jun
    Ng, Wing W. Y.
    Tsang, Eric C. C.
    Zhu, Tao
    Sun, Binbin
    Yeung, Daniel S.
    PROCEEDINGS OF 2008 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2008, : 454 - +
  • [5] Conditional Mutual Information Estimation for Mixed, Discrete and Continuous Data
    Mesner, Octavio Cesar
    Shalizi, Cosma Rohilla
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2021, 67 (01) : 464 - 484
  • [6] Robust Generalization via f Mutual Information
    Esposito, Amedeo Roberto
    Gastpar, Michael
    Issa, Ibrahim
    2020 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2020, : 2723 - 2728
  • [7] A bound on mutual information for image registration
    Skouson, MB
    Guo, QJ
    Liang, ZP
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2001, 20 (08) : 843 - 846
  • [8] Neural Estimators for Conditional Mutual Information Using Nearest Neighbors Sampling
    Molavipour, Sina
    Bassi, German
    Skoglund, Mikael
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 766 - 780
  • [9] Tight Lower Bound of Generalization Error in Ensemble Learning
    Uchida, Masato
    2014 JOINT 7TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS (SCIS) AND 15TH INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS (ISIS), 2014, : 1130 - 1133
  • [10] Feature selection based on conditional mutual information: minimum conditional relevance and minimum conditional redundancy
    HongFang Zhou
    Yao Zhang
    YingJie Zhang
    HongJiang Liu
    Applied Intelligence, 2019, 49 : 883 - 896