Stein Variational Gradient Descent with Variance Reduction

被引:0
|
作者
Nhan Dam [1 ]
Trung Le [1 ]
Viet Huynh [1 ]
Dinh Phung [1 ]
机构
[1] Monash Univ, Clayton, Vic, Australia
来源
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2020年
基金
澳大利亚研究理事会;
关键词
Bayesian inference; variance reduction; statistical machine learning;
D O I
10.1109/ijcnn48605.2020.9206718
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Probabilistic inference is a common and important task in statistical machine learning. The recently proposed Stein variational gradient descent (SVGD) is a generic Bayesian inference method that has been shown to be successfully applied in a wide range of contexts, especially in dealing with large datasets, where existing probabilistic inference methods have been known to be ineffective. In a large-scale data setting, SVGD employs the mini-batch strategy but its mini-batch estimator has large variance, hence compromising its estimation quality in practice. To this end, we propose in this paper a generic SVGD-based inference method that can significantly reduce the variance of mini-batch estimator when working with large datasets. Our experiments on 14 datasets show that the proposed method enjoys substantial and consistent improvements compared with baseline methods in binary classification task and its pseudo-online learning setting, and regression task. Furthermore, our framework is generic and applicable to a wide range of probabilistic inference problems such as in Bayesian neural networks and Markov random fields.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Stein Variational Gradient Descent as Gradient Flow
    Liu, Qiang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [2] Stein Variational Gradient Descent Without Gradient
    Han, Jun
    Liu, Qiang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [3] On the geometry of Stein variational gradient descent
    Duncan, A.
    Nusken, N.
    Szpruch, L.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [4] Projected Stein Variational Gradient Descent
    Chen, Peng
    Ghattas, Omar
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [5] Grassmann Stein Variational Gradient Descent
    Liu, Xing
    Zhu, Harrison
    Ton, Jean-Francois
    Wynne, George
    Duncan, Andrew
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [6] Stein Variational Gradient Descent with Multiple Kernels
    Qingzhong Ai
    Shiyu Liu
    Lirong He
    Zenglin Xu
    Cognitive Computation, 2023, 15 : 672 - 682
  • [7] Stein Variational Gradient Descent as Moment Matching
    Liu, Qiang
    Wang, Dilin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [8] Stein Variational Gradient Descent with Multiple Kernels
    Ai, Qingzhong
    Liu, Shiyu
    He, Lirong
    Xu, Zenglin
    COGNITIVE COMPUTATION, 2023, 15 (02) : 672 - 682
  • [9] Message Passing Stein Variational Gradient Descent
    Zhuo, Jingwei
    Liu, Chang
    Shi, Jiaxin
    Zhu, Jun
    Chen, Ning
    Zhang, Bo
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [10] Stein variational gradient descent with local approximations
    Yan, Liang
    Zhou, Tao
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2021, 386