Deep Recursive Embedding for High-Dimensional Data

被引:8
作者
Zhou, Zixia [1 ]
Zu, Xinrui [2 ]
Wang, Yuanyuan [1 ,3 ]
Lelieveldt, Boudewijn P. F. [4 ]
Tao, Qian [5 ]
机构
[1] Fudan Univ, Dept Elect Engn, Shanghai 200433, Peoples R China
[2] Univ Twente, Fac Elect Engn Math & Comp Sci EEMCS, NL-7522 NB Enschede, Netherlands
[3] Key Lab Med Imaging Comp & Comp Assisted Interven, Shanghai 200032, Peoples R China
[4] Leiden Univ Med Ctr, Dept Radiol, Div Image Proc, NL-2333 ZA Leiden, Netherlands
[5] Delft Univ Technol, Dept Imaging Phys, NL-2628 CJ Delft, Netherlands
关键词
Data visualization; Feature extraction; Training; Manifolds; Unsupervised learning; Standards; Tools; t-distributed stochastic neighbor embedding; uniform manifold approximation and projection; deep embedding network; deep recursive embedding; unsupervised learning; VISUALIZATION;
D O I
10.1109/TVCG.2021.3122388
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Embedding high-dimensional data onto a low-dimensional manifold is of both theoretical and practical value. In this article, we propose to combine deep neural networks (DNN) with mathematics-guided embedding rules for high-dimensional data embedding. We introduce a generic deep embedding network (DEN) framework, which is able to learn a parametric mapping from high-dimensional space to low-dimensional space, guided by well-established objectives such as Kullback-Leibler (KL) divergence minimization. We further propose a recursive strategy, called deep recursive embedding (DRE), to make use of the latent data representations for boosted embedding performance. We exemplify the flexibility of DRE by different architectures and loss functions, and benchmarked our method against the two most popular embedding methods, namely, t-distributed stochastic neighbor embedding (t-SNE) and uniform manifold approximation and projection (UMAP). The proposed DRE method can map out-of-sample data and scale to extremely large datasets. Experiments on a range of public datasets demonstrated improved embedding performance in terms of local and global structure preservation, compared with other state-of-the-art embedding methods. Code is available at https://github.com/tao-aimi/DeepRecursiveEmbedding.
引用
收藏
页码:1237 / 1248
页数:12
相关论文
共 36 条
[1]  
[Anonymous], 2017, ARXIV170807747
[2]  
Balasubramanian M, 2002, SCIENCE, V295
[3]  
Bengio Yoshua, 2013, Statistical Language and Speech Processing. First International Conference, SLSP 2013. Proceedings: LNCS 7978, P1, DOI 10.1007/978-3-642-39593-2_1
[4]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[5]   Data visualization with multidimensional scaling [J].
Buja, Andreas ;
Swayne, Deborah F. ;
Littman, Michael L. ;
Dean, Nathaniel ;
Hofmann, Heike ;
Chen, Lisha .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2008, 17 (02) :444-472
[6]  
Chan DM, 2018, INT SYM COMP ARCHIT, P330, DOI [10.1109/SBAC-PAD.2018.00060, 10.1109/CAHPC.2018.8645912]
[7]  
Elthakeb AT, 2019, 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019)
[8]   Deep learning multidimensional projections [J].
Espadoto, Mateus ;
Tomita Hirata, Nina Sumiko ;
Telea, Alexandru C. .
INFORMATION VISUALIZATION, 2020, 19 (03) :247-269
[9]   Toward a Quantitative Survey of Dimension Reduction Techniques [J].
Espadoto, Mateus ;
Martins, Rafael M. ;
Kerren, Andreas ;
Hirata, Nina S. T. ;
Telea, Alexandru C. .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2021, 27 (03) :2153-2173
[10]   Reducing the dimensionality of data with neural networks [J].
Hinton, G. E. ;
Salakhutdinov, R. R. .
SCIENCE, 2006, 313 (5786) :504-507