Home  | Publications | LBS+21

Dip-Based Deep Embedded Clustering With K-Estimation

MCML Authors

Christian Böhm

Prof. Dr.

Principal Investigator

* Former Principal Investigator

Abstract

The combination of clustering with Deep Learning has gained much attention in recent years. Unsupervised neural networks like autoencoders can autonomously learn the essential structures in a data set. This idea can be combined with clustering objectives to learn relevant features automatically. Unfortunately, they are often based on a k-means framework, from which they inherit various assumptions, like spherical-shaped clusters. Another assumption, also found in approaches outside the k-means-family, is knowing the number of clusters a-priori. In this paper, we present the novel clustering algorithm DipDECK, which can estimate the number of clusters simultaneously to improving a Deep Learning-based clustering objective. Additionally, we can cluster complex data sets without assuming only spherically shaped clusters. Our algorithm works by heavily overestimating the number of clusters in the embedded space of an autoencoder and, based on Hartigan's Dip-test - a statistical test for unimodality - analyses the resulting micro-clusters to determine which to merge. We show in extensive experiments the various benefits of our method: (1) we achieve competitive results while learning the clustering-friendly representation and number of clusters simultaneously; (2) our method is robust regarding parameters, stable in performance, and allows for more flexibility in the cluster shape; (3) we outperform relevant competitors in the estimation of the number of clusters.

inproceedings


KDD 2021

27th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Singapore, Aug 14-18, 2021.
Conference logo
A* Conference

Authors

C. Leiber • L. G. M. Bauer • B. Schelling • C. Böhm • C. Plant

Links

DOI

Research Area

 A3 | Computational Models

BibTeXKey: LBS+21

Back to Top