Individually Conditional Individual Mutual Information Bound on Generalization Error

被引:8
作者
Zhou, Ruida [1 ]
Tian, Chao [1 ]
Liu, Tie [1 ]
机构
[1] Texas A&M Univ, Dept Elect & Comp Engn, College Stn, TX 77843 USA
基金
美国国家科学基金会;
关键词
Mutual information; Training; Random variables; Heuristic algorithms; Training data; Noise measurement; Upper bound; Information-theoretic bounds; generalization error; stochastic gradient Langevin dynamics;
D O I
10.1109/TIT.2022.3144615
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose an information-theoretic bound on the generalization error based on a combination of the error decomposition technique of Bu et al. and the conditional mutual information (CMI) construction of Steinke and Zakynthinou. In a previous work, Haghifam et al. proposed a different bound combining the two aforementioned techniques, which we refer to as the conditional individual mutual information (CIMI) bound. However, in a simple Gaussian setting, both the CMI and the CIMI bounds are order-wise worse than that by Bu et al. This observation motivated us to propose the bound, which overcomes this issue by reducing the conditioning terms in the conditional mutual information. In the process of establishing this bound, a conditional decoupling lemma is established, which also leads to a meaningful dichotomy and comparison among these information-theoretic bounds. As an application of the proposed bound, we analyze the noisy and iterative stochastic gradient Langevin dynamics and provide an upper bound on its generalization error.
引用
收藏
页码:3304 / 3316
页数:13
相关论文
共 50 条
  • [31] On the effect of receiver estimation error upon channel mutual information
    Vosoughi, A
    Scaglione, A
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2006, 54 (02) : 459 - 472
  • [32] Self-Organized Mutual Information Maximization Learning for Improved Generalization Performance
    Kamimura, Ryotaro
    2015 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2015): BIG DATA ANALYTICS FOR HUMAN-CENTRIC SYSTEMS, 2015, : 1613 - 1618
  • [33] Tsallis conditional mutual information in investigating long range correlation in symbol sequences
    Papapetrou, M.
    Kugiumtzis, D.
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2020, 540
  • [34] One Intrusion Detection Method Based On Uniformed Conditional Dynamic Mutual Information
    Lu, Liangfu
    Zhu, Xinhe
    Zhang, Xuyun
    Liu, Junhan
    Bhuiyan, Md Zakirul Alam
    Cui, Guangtai
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (IEEE TRUSTCOM) / 12TH IEEE INTERNATIONAL CONFERENCE ON BIG DATA SCIENCE AND ENGINEERING (IEEE BIGDATASE), 2018, : 1236 - 1241
  • [35] A Rough Set Algorithm for Attribute Reduction via Mutual Information and Conditional Entropy
    Tian, Jing
    Wang, Quan
    Yu, Bing
    Yu, Dan
    2013 10TH INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS AND KNOWLEDGE DISCOVERY (FSKD), 2013, : 567 - 571
  • [36] An upper bound on the ergodic mutual information in Rician fading MIMO channels
    Salo, Jari
    Mikas, Filip
    Vainikainen, Pertti
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2006, 5 (06) : 1415 - 1421
  • [37] An Analysis of the Mutual Information Upper Bound for Sensor-Subset Selection
    Leroy, Idyano
    Petetin, Yohan
    Saucan, Augustin A.
    Clark, Daniel
    2024 27TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, FUSION 2024, 2024,
  • [38] Information-Theoretic Bounds on the Moments of the Generalization Error of Learning Algorithms
    Aminian, Gholamali
    Toni, Laura
    Rodrigues, Miguel R. D.
    2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 682 - 687
  • [39] Mutual information and minimum mean-square error in Gaussian channels
    Guo, DN
    Shamai, S
    Verdú, S
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2005, 51 (04) : 1261 - 1282
  • [40] On mutual information, likelihood ratios, and estimation error for the additive Gaussian channel
    Zakai, M
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2005, 51 (09) : 3017 - 3024