skip to main content
10.1145/3447548.3467316acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Open Access

Dip-based Deep Embedded Clustering with k-Estimation

Published:14 August 2021Publication History

ABSTRACT

The combination of clustering with Deep Learning has gained much attention in recent years. Unsupervised neural networks like autoencoders can autonomously learn the essential structures in a data set. This idea can be combined with clustering objectives to learn relevant features automatically. Unfortunately, they are often based on a k-means framework, from which they inherit various assumptions, like spherical-shaped clusters. Another assumption, also found in approaches outside the k-means-family, is knowing the number of clusters a-priori. In this paper, we present the novel clustering algorithm DipDECK, which can estimate the number of clusters simultaneously to improving a Deep Learning-based clustering objective. Additionally, we can cluster complex data sets without assuming only spherically shaped clusters. Our algorithm works by heavily overestimating the number of clusters in the embedded space of an autoencoder and, based on Hartigan's Dip-test - a statistical test for unimodality - analyses the resulting micro-clusters to determine which to merge. We show in extensive experiments the various benefits of our method: (1) we achieve competitive results while learning the clustering-friendly representation and number of clusters simultaneously; (2) our method is robust regarding parameters, stable in performance, and allows for more flexibility in the cluster shape; (3) we outperform relevant competitors in the estimation of the number of clusters.

Skip Supplemental Material Section

Supplemental Material

KDD21-rst2120.mp4

mp4

118.7 MB

References

  1. Horst Bischof, Alevs Leonardis, and Alexander Selb. 1999. MDL principle for robust vector quantisation. Pattern Analysis & Applications, Vol. 2, 1 (1999), 59--72.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Christian Böhm, Christos Faloutsos, Jia-Yu Pan, and Claudia Plant. 2006. Robust information-theoretic clustering. In SIGKDD. 65--75.Google ScholarGoogle Scholar
  3. Theofilos Chamalis and Aristidis Likas. 2018. The Projected Dip-Means Clustering Algorithm. In Proceedings of the 10th Hellenic Conference on Artificial Intelligence.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Tarin Clanuwat, Mikel Bober-Irizar, Asanobu Kitamoto, Alex Lamb, Kazuaki Yamamoto, and David Ha. 2018. Deep learning for classical Japanese literature. arXiv preprint arXiv:1812.01718 (2018).Google ScholarGoogle Scholar
  5. Dheeru Dua and Casey Graff. 2017. UCI Machine Learning Repository. http://archive.ics.uci.edu/mlGoogle ScholarGoogle Scholar
  6. L. Duan, C. Aggarwal, S. Ma, and S. Sathe. 2019. Improving Spectral Clustering with Deep Embedding and Cluster Estimation. In ICDM. 170--179.Google ScholarGoogle Scholar
  7. Martin Ester, Hans-Peter Kriegel, Jörg Sander, Xiaowei Xu, et almbox. 1996. A density-based algorithm for discovering clusters in large spatial databases with noise.. In SIGKDD, Vol. 96. 226--231.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Yu Feng and Greg Hamerly. 2007. PG-means: learning the number of clusters in data. In Advances in neural information processing systems. 393--400.Google ScholarGoogle Scholar
  9. Kamran Ghasedi Dizaji, Amirhossein Herandi, Cheng Deng, Weidong Cai, and Heng Huang. 2017. Deep clustering via joint convolutional autoencoder embedding and relative entropy minimization. In Proceedings of the IEEE international conference on computer vision. 5736--5745.Google ScholarGoogle ScholarCross RefCross Ref
  10. Xifeng Guo, Long Gao, Xinwang Liu, and Jianping Yin. 2017. Improved Deep Embedded Clustering with Local Structure Preservation. In IJCAI.Google ScholarGoogle Scholar
  11. Greg Hamerly and Charles Elkan. 2004. Learning the k in k-means. In Advances in neural information processing systems. 281--288.Google ScholarGoogle Scholar
  12. J. A. Hartigan and P. M. Hartigan. 1985. The Dip Test of Unimodality. Ann. Statist., Vol. 13, 1 (03 1985), 70--84. https://doi.org/10.1214/aos/1176346577Google ScholarGoogle Scholar
  13. Sebastian Houben, Johannes Stallkamp, Jan Salmen, Marc Schlipsing, and Christian Igel. 2013. Detection of traffic signs in real-world images: The German Traffic Sign Detection Benchmark. In IJCNN. IEEE, 1--8.Google ScholarGoogle Scholar
  14. Jonathan J. Hull. 1994. A database for handwritten text recognition research. IEEE Transactions on pattern analysis and machine intelligence, Vol. 16, 5 (1994), 550--554.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Zhuxi Jiang, Yin Zheng, Huachun Tan, Bangsheng Tang, and Hanning Zhou. 2017. Variational Deep Embedding: An Unsupervised and Generative Approach to Clustering. In IJCAI. 1965--1972. https://doi.org/10.24963/ijcai.2017/273Google ScholarGoogle Scholar
  16. Argyris Kalogeratos and Aristidis Likas. 2012. Dip-means: an incremental clustering method for estimating the number of clusters. In Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  17. Yann Lecun. 1987. PhD thesis: Modeles connexionnistes de l'apprentissage (connectionist learning models). Universite P. et M. Curie (Paris 6).Google ScholarGoogle Scholar
  18. Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. 1998. Gradient-based learning applied to document recognition. Proc. IEEE, Vol. 86, 11 (1998), 2278--2324.Google ScholarGoogle ScholarCross RefCross Ref
  19. Samuel Maurus and Claudia Plant. 2016. Skinny-dip: clustering in a sea of noise. In SIGKDD. 1055--1064.Google ScholarGoogle Scholar
  20. Dominik Mautz, Claudia Plant, and Christian Böhm. 2019. Deep embedded cluster tree. In ICDM. IEEE, 1258--1263.Google ScholarGoogle Scholar
  21. Tom Monnier, Thibault Groueix, and Mathieu Aubry. 2020. Deep Transformation-Invariant Clustering. In NeurIPS.Google ScholarGoogle Scholar
  22. Sudipto Mukherjee, Himanshu Asnani, Eugene Lin, and Sreeram Kannan. 2019. ClusterGAN: Latent Space Clustering in Generative Adversarial Networks. In AAAI. 4610--4617.Google ScholarGoogle Scholar
  23. Dan Pelleg and Andrew W. Moore. 2000. X-Means: Extending K-Means with Efficient Estimation of the Number of Clusters. In ICML (ICML '00).Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. B. Schelling, L. Bauer, S. Behzadi Soheil, and C. Plant. 2020. Utilizing Structure-rich Features to improve Clustering. In ECML-PKDD 2020.Google ScholarGoogle Scholar
  25. B. Schelling and C. Plant. 2018. Dip Transformation: Enhancing the Structure of a Dataset and Thereby Improving Clustering. In ICDM.Google ScholarGoogle Scholar
  26. Nguyen Xuan Vinh, Julien Epps, and James Bailey. 2010. Information Theoretic Measures for Clusterings Comparison: Variants, Properties, Normalization and Correction for Chance. Journal of Machine Learning Research (2010).Google ScholarGoogle Scholar
  27. Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms. [arXiv]cs.LG/1708.07747 [cs.LG]Google ScholarGoogle Scholar
  28. Junyuan Xie, Ross B. Girshick, and Ali Farhadi. 2016. Unsupervised Deep Embedding for Clustering Analysis. In ICML (JMLR Workshop and Conference Proceedings).Google ScholarGoogle Scholar
  29. Bo Yang, Xiao Fu, Nicholas D. Sidiropoulos, and Mingyi Hong. 2017. Towards K-means-friendly Spaces: Simultaneous Deep Learning and Clustering. In ICML.Google ScholarGoogle Scholar
  30. Jianwei Yang, Devi Parikh, and Dhruv Batra. 2016. Joint Unsupervised Learning of Deep Representations and Image Clusters. In CVPR.Google ScholarGoogle Scholar
  31. Lihi Zelnik-Manor and Pietro Perona. 2005. Self-Tuning Spectral Clustering. In Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar

Index Terms

  1. Dip-based Deep Embedded Clustering with k-Estimation

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            KDD '21: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining
            August 2021
            4259 pages
            ISBN:9781450383325
            DOI:10.1145/3447548

            Copyright © 2021 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 14 August 2021

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            Overall Acceptance Rate1,133of8,635submissions,13%

            Upcoming Conference

            KDD '24

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader