Bottlenecks CLUB: Unifying Information-Theoretic Trade-Offs Among Complexity, Leakage, and Utility

被引:7
作者
Razeghi, Behrooz [1 ]
Calmon, Flavio P. [2 ]
Gunduz, Deniz [3 ]
Voloshynovskiy, Slava [1 ]
机构
[1] Univ Geneva, Dept Comp Sci, CH-1227 Geneva, Switzerland
[2] Harvard Univ, Sch Engn & Appl Sci, Cambridge, MA 02134 USA
[3] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2BT, England
基金
瑞士国家科学基金会; 英国工程与自然科学研究理事会;
关键词
Training; Privacy; Machine learning algorithms; Neural networks; Generative adversarial networks; Mathematical models; Loss measurement; Information-theoretic privacy; statistical inference; information bottleneck; obfuscation; generative models;
D O I
10.1109/TIFS.2023.3262112
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Bottleneck problems are an important class of optimization problems that have recently gained increasing attention in the domain of machine learning and information theory. They are widely used in generative models, fair machine learning algorithms, design of privacy-assuring mechanisms, and appear as information-theoretic performance bounds in various multi-user communication problems. In this work, we propose a general family of optimization problems, termed as complexity-leakage-utility bottleneck (CLUB) model, which (i) provides a unified theoretical framework that generalizes most of the state-of-the-art literature for the information-theoretic privacy models, (ii) establishes a new interpretation of the popular generative and discriminative models, (iii) constructs new insights for the generative compression models, and (iv) can be used to obtain fair generative models. We first formulate the CLUB model as a complexity-constrained privacy-utility optimization problem. We then connect it with the closely related bottleneck problems, namely information bottleneck (IB), privacy funnel (PF), deterministic IB (DIB), conditional entropy bottleneck (CEB), and conditional PF (CPF). We show that the CLUB model generalizes all these problems as well as most other information-theoretic privacy models. Then, we construct the deep variational CLUB (DVCLUB) models by employing neural networks to parameterize variational approximations of the associated information quantities. Building upon these information quantities, we present unified objectives of the supervised and unsupervised DVCLUB models. Leveraging the DVCLUB model in an unsupervised setup, we then connect it with state-of-the-art generative models, such as variational auto-encoders (VAEs), generative adversarial networks (GANs), as well as the Wasserstein GAN (WGAN), Wasserstein auto-encoder (WAE), and adversarial auto-encoder (AAE) models through the optimal transport (OT) problem. We then show that the DVCLUB model can also be used in fair representation learning problems, where the goal is to mitigate the undesired bias during the training phase of a machine learning model. We conduct extensive quantitative experiments on colored-MNIST and CelebA datasets.
引用
收藏
页码:2060 / 2075
页数:16
相关论文
共 75 条
  • [1] Abadi M, 2016, P OSDI, P265, DOI DOI 10.1016/0076-6879(83)01039-3
  • [2] Alghamdi W., 2022, PROC NEURIPS
  • [3] Entropy-based distortion measure for image coding
    Andre, Thomas
    Antonini, Marc
    Barlaud, Michel
    Gray, Robert M.
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP 2006, PROCEEDINGS, 2006, : 1157 - +
  • [4] Arimoto S, 1977, TOPICS INFORM THEORY
  • [5] Arjovsky M, 2017, PR MACH LEARN RES, V70
  • [6] Estimation Efficiency Under Privacy Constraints
    Asoodeh, Shahab
    Diaz, Mario
    Alajaji, Fady
    Linder, Tamas
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2019, 65 (03) : 1512 - 1534
  • [7] Information Extraction Under Privacy Constraints
    Asoodeh, Shahab
    Diaz, Mario
    Alajaji, Fady
    Linder, Tamas
    [J]. INFORMATION, 2016, 7 (01)
  • [8] Atashin Amir Ahooye, 2021, WiseML '21: Proceedings of the 3rd ACM Workshop on Wireless Security and Machine Learning, P91, DOI 10.1145/3468218.3469040
  • [9] Basciftci YO, 2016, 2016 INFORMATION THEORY AND APPLICATIONS WORKSHOP (ITA)
  • [10] Bauer M, 2019, PR MACH LEARN RES, V89, P66