Abstract
Manifold is considered to be the explicit form of data, so the smoothness of manifold is related to data dimensionality. Data becomes sparse in the high-dimensional space, which hardly affords sufficient information. Thus, it is a challenge for smooth manifold extraction from the data existing in high-dimensional space. To address this issue, here proposes a deep model of having three-hidden layers for smooth manifold extraction. Our thought is originate from the view of the optimal transportation mass theory. Because high-dimensional data resides around a low-dimensional manifold, we can reconstruct a lower dimensional manifold in high-dimensional space. To guarantee the quality of the reconstructed manifold, the sampling condition is used in order to reconstruct a discrete surface that can converge to an original surface. Meanwhile, the loss function derived by Brenier theorem minimizes the error between the original data distribution and the reconstructed data distribution. In addition, to promote the generalization ability of our model, the neurons in the hidden layers are turned off with the probability manner just during training. Experimental results show our method outperforms the state-of-the-art methods in smooth manifold extraction. We find that as for a deep model, the manner of turning off some neurons using probability carries more weights in improving the smoothness of manifold extraction than investing the effort of simply stacking hidden layers. Moreover, the manner of turning off some neurons using probability also mitigates over-fitting to a certain extent. Our finding also suggests that for high-dimensional space, the results of manifold extraction using the model possessing a deep architecture basic paradigm are superior to that of using the state-of-the-art methods.
Similar content being viewed by others
Data availability
All dataset can be found at http://archive.ics.uci.edu/ml/.
References
Andras P (2015) High-dimensional function approximation using local linear embedding. In: International Joint Conference on Neural Networks (IJCNN), pp 1–8
Andras P (2018) High-dimensional function approximation with neural networks for large volumes of data. IEEE Trans Neural Netw Learning Syst 29(2):500–508
Brenier Y (1991) Polar factorization and monotone rearrangement of vector-valued functions. Commun Pure Appl Math 44(4):375–417
Dai J, Hu H, Hu Q (2018) Locally Linear Approximation Approach for Incomplete Data. IEEE Trans Cybern 48(6):1720–1732
Deng L, Hinton G, Kingsbury B (2013) New types of deep neural network learning for speech recognition and related applications: an overview. In: 2013 IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pp 8599–8603
Deng Y, Bao F, Dai Q, Wu LF (2019) Scalable analysis of cell-type composition from single-cell transcriptomics using deep recurrent learning. Nat Methods 16(4):311–314
Diederik PK, Jimmy LB (2015) Adam: a method for stochastic optimization. arXiv:1412.6980
Gheisari M, Guojun Wang Md, Bhuiyan ZA (2017) A survey on deep learning in big data. CSE/EUC 2:173–180
Goldberg Y, Zakai A, Kushnir D, Ritov Y (2018) Manifold learning: the price of normalization. J Mach Learn Res 9:1909–1939
Gong J, Ma H, Teng Z, Teng Qi, Zhang H, Linfeng Du, Shuai Chen Md, Bhuiyan ZA, Li J, Liu M (2020) Hierarchical graph transformer-based deep learning model for large-scale multi-label text classification. IEEE Access 8:30885–30896
Gu X, Luo F, Sun J, Yau S-T (2016) Variational principles for minkowski type problems, discrete optimal transport and discrete monge-ampere equations. Asian J Math 20(2):383–398
Guo X, Minai AA, Lu LJ (2017) Feature selection using multiple auto-encoders. In: IEEE 2017 International Joint Conference on Neural Networks (IJCNN), pp 4602–4609
He X, Niyogi P (2004) Locality preserving projections. In: Advances in neural information processing systems. Neural processing letters, pp 153–160
Jake L, Martin K, Naomi A (2017) Principal component analysis. Nat Methods 14(7):641–642
Le Cun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444
Liu L, Yu M, Shao L (2016) Unsupervised local feature hashing for image similarity search. IEEE Trans Cybern 46(11):2548–2558
Liu Y, Li Z, Zhou C, Jiang Y, Sun J, Wang M, He X (2019) Generative adversarial active learning for unsupervised outlier detection. IEEE Trans Knowl Data Eng. https://doi.org/10.1109/TKDE.2019.2905606
Ma J, Yu MK, Fong S, Ono K, Sage E (2018) Using deep learning to model the hierarchical structure and function of a cell. Nat Methods 15:290–298
Shihavuddin A, Basu S, Rexhepaj E, Delestro F, Menezes N (2017) Smooth 2D manifold extraction from 3D image stack. Nat Commun 8:1–8
Tang J, Shao L, Li X, Lu K (2016) A local structural descriptor for image matching via normalized graph laplacian embedding. IEEE Trans Cybern 46(2):410–420
Tompson J, Stein M, Le Cun Y, Perlin K (2014) Real-time continuous pose recovery of human hands using convolutional networks. ACM Trans Graphics 33:169–169
Vander Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9:2579–2605
Villani C (2003) Topics in optimal transportation. graduate studies in mathematics. American Mathematical Society, Providence 58
Villani C (2008) Optimal Transport: Old and New. Springer Science & Business Media, 338
Shulin Wang, Fang Chen (2015) Spectral Clustering of High-dimensional Data via Nonnegative Matrix Factorization. In 2015 International Joint Conference on Neural Networks (IJCNN), pp 1–1
Xiong L, Kui Xu, Tian K, Shao Y, Tang L (2019) SCALE method for single-cell ATAC-seq analysis via latent feature extraction. Nat Commun 10:1–10
Zhang Z, Zha H (2004) Principal manifolds and nonlinear dimensionality reduction via tangent space alignment. SIAM J Sci Comput 26(1):313–338
Zhang J, Jun Y, Dacheng T (2018) Local deep-feature alignment for unsupervised dimension reduction. IEEE Trans Image Process 27(5):2420–2432
Zheng J, Wang J, Chen Y, Chen S, Chen J, Zhong W, Wu W (2021a) Effective approximation of high-dimensional space using neural networks. J Supercomputing 78:4377–4397
Zheng J, Wang J, Chen Y, Chen S, Chen J, Zhong W, Wenling Wu (2021b) Neural networks trained with high-dimensional functions approximation data in high-dimensional space. J Intell Fuzzy Sys 41:3739–3750
Funding
This research was supported by the Science and Technology Research Program of Chongqing Municipal Education Commission of China (Grant No KJQN201903003).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zheng, J. Smooth manifold extraction in high-dimensional data using a deep model. J Ambient Intell Human Comput 13, 4467–4476 (2022). https://doi.org/10.1007/s12652-022-03824-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12652-022-03824-x