Isometric Quotient Variational Auto-Encoders for Structure-Preserving Representation Learning

被引:0
作者
Huh, In [1 ]
Jeong, Changwook [2 ]
Choe, Jae Myung [1 ]
Kim, Young-Gu [1 ]
Kim, Dae Sin [1 ]
机构
[1] Samsung Elect, Innovat Ctr, CSE Team, Suwon, South Korea
[2] UNIST, Grad Sch Semicond Mat & Devices Engn, Ulsan, South Korea
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study structure-preserving low-dimensional representation of a data manifold embedded in a high-dimensional observation space based on variational auto-encoders (VAEs). We approach this by decomposing the data manifold M as M= M/G x G, where G and M/G are a group of symmetry transformations and a quotient space ofMup to G, respectively. From this perspective, we define the structure-preserving representation of such a manifold as a latent space Z which is isometrically isomorphic (i.e., distance-preserving) to the quotient space M/G rather M (i.e., symmetry-preserving). To this end, we propose a novel auto-encoding framework, named isometric quotient VAEs (IQVAEs), that can extract the quotient space from observations and learn the Riemannian isometry of the extracted quotient in an unsupervised manner. Empirical proof-of-concept experiments reveal that the proposed method can find a meaningful representation of the learned data and outperform other competitors for downstream tasks.
引用
收藏
页数:13
相关论文
共 46 条
[1]  
[Anonymous], 2020, PMLR
[2]  
[Anonymous], 2016, PMLR
[3]  
Arvanitidis G., 2018, P INT C LEARN REPR
[4]  
Arvanitidis G, 2021, PR MACH LEARN RES, V130, P631
[5]  
BEC F, 2020, INT C LEARN REPR, V40, P1937
[6]  
Benton G., 2020, Advances in Neural Information Processing Systems, V33, P17605
[7]   Why Deep Learning Works: A Manifold Disentanglement Perspective [J].
Brahma, Pratik Prabhanjan ;
Wu, Dapeng ;
She, Yiyuan .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (10) :1997-2008
[8]  
Chadebec Clement, 2022, ARXIV220907370
[9]  
Chen Nutan, 2020, P MACHINE LEARNING R, P1587
[10]  
Falorsi Luca, 2021, ICML 2018 WORKSH THE