Learning to Generate with Memory

被引:0
|
作者
Li, Chongxuan [1 ]
Zhu, Jun [1 ]
Zhang, Bo [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Tech, State Key Lab Intell Tech & Sys, TNList Lab,Ctr Bioinspired Comp Res, Beijing 100084, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Memory units have been widely used to enrich the capabilities of deep networks on capturing long-term dependencies in reasoning and prediction tasks, but little investigation exists on deep generative models (DGMs) which are good at inferring high-level invariant representations from unlabeled data. This paper presents a deep generative model with a possibly large external memory and an attention mechanism to capture the local detail information that is often lost in the bottom-up abstraction process in representation learning. By adopting a smooth attention model, the whole network is trained end-to-end by optimizing a variational bound of data likelihood via auto-encoding variational Bayesian methods, where an asymmetric recognition network is learnt jointly to infer high-level invariant representations. The asymmetric architecture can reduce the competition between bottom-up invariant feature extraction and top-down generation of instance details. Our experiments on several datasets demonstrate that memory can significantly boost the performance of DGMs on various tasks, including density estimation, image generation, and missing value imputation, and DGMs with memory can achieve state-of-the-art quantitative results.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Learning can generate long memory
    Chevillon, Guillaume
    Mavroeidis, Sophocles
    JOURNAL OF ECONOMETRICS, 2017, 198 (01) : 1 - 9
  • [2] Competitive learning to generate sparse representations for associative memory
    Sacouto, Luis
    Wichert, Andreas
    NEURAL NETWORKS, 2023, 168 : 32 - 43
  • [3] Learning to Generate Questions by Learning What not to Generate
    Liu, Bang
    Zhao, Mingjun
    Niu, Di
    Lai, Kunfeng
    He, Yancheng
    Wei, Haojie
    Xu, Yu
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 1106 - 1118
  • [4] Memory Replay GANs: learning to generate images from new categories without forgetting
    Wu, Chenshen
    Herranz, Luis
    Liu, Xialei
    Wang, Yaxing
    van de Weijer, Joost
    Raducanu, Bogdan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] Learning to Prove Theorems by Learning to Generate Theorems
    Wang, Mingzhe
    Deng, Jia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Learning to Generate and Edit Hairstyles
    Yin, Weidong
    Fu, Yanwei
    Ma, Yiqing
    Jiang, Yu-Gang
    Xiang, Tao
    Xue, Xiangyang
    PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 1627 - 1635
  • [7] Learning to Generate Wasserstein Barycenters
    Lacombe, Julien
    Digne, Julie
    Courty, Nicolas
    Bonneel, Nicolas
    JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2023, 65 (02) : 354 - 370
  • [8] Learning to Generate Popular Headlines
    Omidvar, Amin
    An, Aijun
    IEEE ACCESS, 2023, 11 : 60904 - 60914
  • [9] Learning to Generate Wasserstein Barycenters
    Julien Lacombe
    Julie Digne
    Nicolas Courty
    Nicolas Bonneel
    Journal of Mathematical Imaging and Vision, 2023, 65 : 354 - 370
  • [10] Learning to Generate Ambiguous Sequences
    Iclanzan, David
    Szilagyi, Laszlo
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 110 - 121