Smooth manifold extraction in high-dimensional data using a deep model

被引:0
作者
Jian Zheng
机构
[1] Chongqing Aerospace Polytechnic,
来源
Journal of Ambient Intelligence and Humanized Computing | 2022年 / 13卷
关键词
Deep model; Smooth manifold; High-dimensional data;
D O I
暂无
中图分类号
学科分类号
摘要
Manifold is considered to be the explicit form of data, so the smoothness of manifold is related to data dimensionality. Data becomes sparse in the high-dimensional space, which hardly affords sufficient information. Thus, it is a challenge for smooth manifold extraction from the data existing in high-dimensional space. To address this issue, here proposes a deep model of having three-hidden layers for smooth manifold extraction. Our thought is originate from the view of the optimal transportation mass theory. Because high-dimensional data resides around a low-dimensional manifold, we can reconstruct a lower dimensional manifold in high-dimensional space. To guarantee the quality of the reconstructed manifold, the sampling condition is used in order to reconstruct a discrete surface that can converge to an original surface. Meanwhile, the loss function derived by Brenier theorem minimizes the error between the original data distribution and the reconstructed data distribution. In addition, to promote the generalization ability of our model, the neurons in the hidden layers are turned off with the probability manner just during training. Experimental results show our method outperforms the state-of-the-art methods in smooth manifold extraction. We find that as for a deep model, the manner of turning off some neurons using probability carries more weights in improving the smoothness of manifold extraction than investing the effort of simply stacking hidden layers. Moreover, the manner of turning off some neurons using probability also mitigates over-fitting to a certain extent. Our finding also suggests that for high-dimensional space, the results of manifold extraction using the model possessing a deep architecture basic paradigm are superior to that of using the state-of-the-art methods.
引用
收藏
页码:4467 / 4476
页数:9
相关论文
共 90 条
[1]  
Andras P(2018)High-dimensional function approximation with neural networks for large volumes of data IEEE Trans Neural Netw Learning Syst 29 500-508
[2]  
Brenier Y(1991)Polar factorization and monotone rearrangement of vector-valued functions Commun Pure Appl Math 44 375-417
[3]  
Dai J(2018)Locally Linear Approximation Approach for Incomplete Data IEEE Trans Cybern 48 1720-1732
[4]  
Hu H(2019)Scalable analysis of cell-type composition from single-cell transcriptomics using deep recurrent learning Nat Methods 16 311-314
[5]  
Hu Q(2017)A survey on deep learning in big data CSE/EUC 2 173-180
[6]  
Deng Y(2018)Manifold learning: the price of normalization J Mach Learn Res 9 1909-1939
[7]  
Bao F(2020)Hierarchical graph transformer-based deep learning model for large-scale multi-label text classification IEEE Access 8 30885-30896
[8]  
Dai Q(2016)Variational principles for minkowski type problems, discrete optimal transport and discrete monge-ampere equations Asian J Math 20 383-398
[9]  
Wu LF(2017)Principal component analysis Nat Methods 14 641-642
[10]  
Gheisari M(2015)Deep learning Nature 521 436-444