Task-Agnostic Amortized Inference of Gaussian Process Hyperparameters

被引:0
|
作者
Liu, Sulin [1 ]
Sun, Xingyuan [1 ]
Ramadge, Peter J. [1 ]
Adams, Ryan P. [1 ]
机构
[1] Princeton Univ, Princeton, NJ 08544 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gaussian processes (GPs) are flexible priors for modeling functions. However, their success depends on the kernel accurately reflecting the properties of the data. One of the appeals of the GP framework is that the marginal likelihood of the kernel hyperparameters is often available in closed form, enabling optimization and sampling procedures to fit these hyperparameters to data. Unfortunately, point-wise evaluation of the marginal likelihood is expensive due to the need to solve a linear system; searching or sampling the space of hyperparameters thus often dominates the practical cost of using GPs. We introduce an approach to the identification of kernel hyperparameters in GP regression and related problems that sidesteps the need for costly marginal likelihoods. Our strategy is to "amortize" inference over hyperparameters by training a single neural network, which consumes a set of regression data and produces an estimate of the kernel function, useful across different tasks. To accommodate the varying dimension and cardinality of different regression problems, we use a hierarchical self-attention-based neural network that produces estimates of the hyperparameters which are invariant to the order of the input data points and data dimensions. We show that a single neural model trained on synthetic data is able to generalize directly to several different unseen real-world GP use cases. Our experiments demonstrate that the estimated hyperparameters are comparable in quality to those from the conventional model selection procedures, while being much faster to obtain, significantly accelerating GP regression and its related applications such as Bayesian optimization and Bayesian quadrature. The code and pre-trained model are available at https://github.com/PrincetonLIPS/AHGP.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] FADE: Fusing the Assets of Decoder and Encoder for Task-Agnostic Upsampling
    Lu, Hao
    Liu, Wenze
    Fu, Hongtao
    Cao, Zhiguo
    COMPUTER VISION - ECCV 2022, PT XXVII, 2022, 13687 : 231 - 247
  • [32] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following
    Ye, Seonghyeon
    Hwang, Hyeonbin
    Yang, Sohee
    Yun, Hyeongu
    Kim, Yireun
    Seo, Minjoon
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19386 - 19394
  • [33] A Self-supervised Task-agnostic Embedding for EEG Signals
    Partovi, Andi
    Burkitt, Anthony N.
    Grayden, David
    2023 11TH INTERNATIONAL IEEE/EMBS CONFERENCE ON NEURAL ENGINEERING, NER, 2023,
  • [34] Task-Agnostic Vision Transformer for Distributed Learning of Image Processing
    Kim, Boah
    Kim, Jeongsol
    Ye, Jong Chul
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 203 - 218
  • [35] Extract then Distill: Efficient and Effective Task-Agnostic BERT Distillation
    Chen, Cheng
    Yin, Yichun
    Shang, Lifeng
    Wang, Zhi
    Jiang, Xin
    Chen, Xiao
    Liu, Qun
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT III, 2021, 12893 : 570 - 581
  • [36] Task-Agnostic Adaptation for Safe Human-Robot Handover
    Liu, Ruixuan
    Chen, Rui
    Liu, Changliu
    IFAC PAPERSONLINE, 2022, 55 (41): : 175 - 180
  • [37] CodePrompt: Task-Agnostic Prefix Tuning for Program and Language Generation
    Choi, YunSeok
    Lee, Jee-Hyong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5282 - 5297
  • [38] Continual deep reinforcement learning with task-agnostic policy distillation
    Hafez, Muhammad Burhan
    Erekmen, Kerim
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [39] Interesting Object, Curious Agent: Learning Task-Agnostic Exploration
    Parisi, Simone
    Dean, Victoria
    Pathak, Deepak
    Gupta, Abhinav
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [40] How Effective is Task-Agnostic Data Augmentation for Pretrained Transformers?
    Longpre, Shayne
    Wang, Yu
    DuBois, Christopher
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4401 - 4411