Deep divergence-based approach to clustering.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

A promising direction in deep learning research consists in learning representations and simultaneously discovering cluster structure in unlabeled data by optimizing a discriminative loss function. As opposed to supervised deep learning, this line of research is in its infancy, and how to design and optimize suitable loss functions to train deep neural networks for clustering is still an open question. Our contribution to this emerging field is a new deep clustering network that leverages the discriminative power of information-theoretic divergence measures, which have been shown to be effective in traditional clustering. We propose a novel loss function that incorporates geometric regularization constraints, thus avoiding degenerate structures of the resulting clustering partition. Experiments on synthetic benchmarks and real datasets show that the proposed network achieves competitive performance with respect to other state-of-the-art methods, scales well to large datasets, and does not require pre-training steps.

Authors

  • Michael Kampffmeyer
    Machine Learning Group, UiT the Arctic University of Norway, Norway (1). Electronic address: michael.c.kampffmeyer@uit.no.
  • Sigurd Løkse
    Machine Learning Group, UiT the Arctic University of Norway, Norway (1).
  • Filippo M Bianchi
    Machine Learning Group, UiT the Arctic University of Norway, Norway (1).
  • Lorenzo Livi
    Department of Computer Science, College of Engineering, Mathematics and Physical Sciences, University of Exeter, Exeter EX4 4QF, United Kingdom.
  • Arnt-Børre Salberg
    Norwegian Computing Center, Oslo, Norway.
  • Robert Jenssen
    Department of Physics and Technology, University of Tromsø - The Arctic University of Norway, Tromsø, Norway; Norwegian Centre for Integrated Care and Telemedicine, University Hospital of North Norway, Tromsø, Norway.