Skip to main content
Log in

Smooth manifold extraction in high-dimensional data using a deep model

  • Original Research
  • Published:
Journal of Ambient Intelligence and Humanized Computing Aims and scope Submit manuscript

Abstract

Manifold is considered to be the explicit form of data, so the smoothness of manifold is related to data dimensionality. Data becomes sparse in the high-dimensional space, which hardly affords sufficient information. Thus, it is a challenge for smooth manifold extraction from the data existing in high-dimensional space. To address this issue, here proposes a deep model of having three-hidden layers for smooth manifold extraction. Our thought is originate from the view of the optimal transportation mass theory. Because high-dimensional data resides around a low-dimensional manifold, we can reconstruct a lower dimensional manifold in high-dimensional space. To guarantee the quality of the reconstructed manifold, the sampling condition is used in order to reconstruct a discrete surface that can converge to an original surface. Meanwhile, the loss function derived by Brenier theorem minimizes the error between the original data distribution and the reconstructed data distribution. In addition, to promote the generalization ability of our model, the neurons in the hidden layers are turned off with the probability manner just during training. Experimental results show our method outperforms the state-of-the-art methods in smooth manifold extraction. We find that as for a deep model, the manner of turning off some neurons using probability carries more weights in improving the smoothness of manifold extraction than investing the effort of simply stacking hidden layers. Moreover, the manner of turning off some neurons using probability also mitigates over-fitting to a certain extent. Our finding also suggests that for high-dimensional space, the results of manifold extraction using the model possessing a deep architecture basic paradigm are superior to that of using the state-of-the-art methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Data availability

All dataset can be found at http://archive.ics.uci.edu/ml/.

References

  • Andras P (2015) High-dimensional function approximation using local linear embedding. In: International Joint Conference on Neural Networks (IJCNN), pp 1–8

  • Andras P (2018) High-dimensional function approximation with neural networks for large volumes of data. IEEE Trans Neural Netw Learning Syst 29(2):500–508

    Article  MathSciNet  Google Scholar 

  • Brenier Y (1991) Polar factorization and monotone rearrangement of vector-valued functions. Commun Pure Appl Math 44(4):375–417

    Article  MathSciNet  Google Scholar 

  • Dai J, Hu H, Hu Q (2018) Locally Linear Approximation Approach for Incomplete Data. IEEE Trans Cybern 48(6):1720–1732

    Article  Google Scholar 

  • Deng L, Hinton G, Kingsbury B (2013) New types of deep neural network learning for speech recognition and related applications: an overview. In: 2013 IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pp 8599–8603

  • Deng Y, Bao F, Dai Q, Wu LF (2019) Scalable analysis of cell-type composition from single-cell transcriptomics using deep recurrent learning. Nat Methods 16(4):311–314

    Article  Google Scholar 

  • Diederik PK, Jimmy LB (2015) Adam: a method for stochastic optimization. arXiv:1412.6980

  • Gheisari M, Guojun Wang Md, Bhuiyan ZA (2017) A survey on deep learning in big data. CSE/EUC 2:173–180

    Google Scholar 

  • Goldberg Y, Zakai A, Kushnir D, Ritov Y (2018) Manifold learning: the price of normalization. J Mach Learn Res 9:1909–1939

    MathSciNet  MATH  Google Scholar 

  • Gong J, Ma H, Teng Z, Teng Qi, Zhang H, Linfeng Du, Shuai Chen Md, Bhuiyan ZA, Li J, Liu M (2020) Hierarchical graph transformer-based deep learning model for large-scale multi-label text classification. IEEE Access 8:30885–30896

    Article  Google Scholar 

  • Gu X, Luo F, Sun J, Yau S-T (2016) Variational principles for minkowski type problems, discrete optimal transport and discrete monge-ampere equations. Asian J Math 20(2):383–398

    Article  MathSciNet  Google Scholar 

  • Guo X, Minai AA, Lu LJ (2017) Feature selection using multiple auto-encoders. In: IEEE 2017 International Joint Conference on Neural Networks (IJCNN), pp 4602–4609

  • He X, Niyogi P (2004) Locality preserving projections. In: Advances in neural information processing systems. Neural processing letters, pp 153–160

  • Jake L, Martin K, Naomi A (2017) Principal component analysis. Nat Methods 14(7):641–642

    Article  Google Scholar 

  • Le Cun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444

    Article  Google Scholar 

  • Liu L, Yu M, Shao L (2016) Unsupervised local feature hashing for image similarity search. IEEE Trans Cybern 46(11):2548–2558

    Article  Google Scholar 

  • Liu Y, Li Z, Zhou C, Jiang Y, Sun J, Wang M, He X (2019) Generative adversarial active learning for unsupervised outlier detection. IEEE Trans Knowl Data Eng. https://doi.org/10.1109/TKDE.2019.2905606

    Article  Google Scholar 

  • Ma J, Yu MK, Fong S, Ono K, Sage E (2018) Using deep learning to model the hierarchical structure and function of a cell. Nat Methods 15:290–298

    Article  Google Scholar 

  • Shihavuddin A, Basu S, Rexhepaj E, Delestro F, Menezes N (2017) Smooth 2D manifold extraction from 3D image stack. Nat Commun 8:1–8

    Article  Google Scholar 

  • Tang J, Shao L, Li X, Lu K (2016) A local structural descriptor for image matching via normalized graph laplacian embedding. IEEE Trans Cybern 46(2):410–420

    Article  Google Scholar 

  • Tompson J, Stein M, Le Cun Y, Perlin K (2014) Real-time continuous pose recovery of human hands using convolutional networks. ACM Trans Graphics 33:169–169

    Article  Google Scholar 

  • Vander Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9:2579–2605

    MATH  Google Scholar 

  • Villani C (2003) Topics in optimal transportation. graduate studies in mathematics. American Mathematical Society, Providence 58

  • Villani C (2008) Optimal Transport: Old and New. Springer Science & Business Media, 338

  • Shulin Wang, Fang Chen (2015) Spectral Clustering of High-dimensional Data via Nonnegative Matrix Factorization. In 2015 International Joint Conference on Neural Networks (IJCNN), pp 1–1

  • Xiong L, Kui Xu, Tian K, Shao Y, Tang L (2019) SCALE method for single-cell ATAC-seq analysis via latent feature extraction. Nat Commun 10:1–10

    Article  Google Scholar 

  • Zhang Z, Zha H (2004) Principal manifolds and nonlinear dimensionality reduction via tangent space alignment. SIAM J Sci Comput 26(1):313–338

    Article  MathSciNet  Google Scholar 

  • Zhang J, Jun Y, Dacheng T (2018) Local deep-feature alignment for unsupervised dimension reduction. IEEE Trans Image Process 27(5):2420–2432

    Article  MathSciNet  Google Scholar 

  • Zheng J, Wang J, Chen Y, Chen S, Chen J, Zhong W, Wu W (2021a) Effective approximation of high-dimensional space using neural networks. J Supercomputing 78:4377–4397

    Article  Google Scholar 

  • Zheng J, Wang J, Chen Y, Chen S, Chen J, Zhong W, Wenling Wu (2021b) Neural networks trained with high-dimensional functions approximation data in high-dimensional space. J Intell Fuzzy Sys 41:3739–3750

    Article  Google Scholar 

Download references

Funding

This research was supported by the Science and Technology Research Program of Chongqing Municipal Education Commission of China (Grant No KJQN201903003).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jian Zheng.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zheng, J. Smooth manifold extraction in high-dimensional data using a deep model. J Ambient Intell Human Comput 13, 4467–4476 (2022). https://doi.org/10.1007/s12652-022-03824-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12652-022-03824-x

Keywords

Navigation