Joint optimization of an autoencoder for clustering and embedding

Research output: Journal contributionsJournal articlesResearchpeer-review

Standard

Joint optimization of an autoencoder for clustering and embedding. / Boubekki, Ahcène; Kampffmeyer, Michael; Brefeld, Ulf et al.
In: Machine Learning, Vol. 110, No. 7, 01.07.2021, p. 1901-1937.

Research output: Journal contributionsJournal articlesResearchpeer-review

Harvard

APA

Vancouver

Boubekki A, Kampffmeyer M, Brefeld U, Jenssen R. Joint optimization of an autoencoder for clustering and embedding. Machine Learning. 2021 Jul 1;110(7):1901-1937. doi: 10.1007/s10994-021-06015-5

Bibtex

@article{5d22629599494a47993f95b18dfd860c,
title = "Joint optimization of an autoencoder for clustering and embedding",
abstract = "Deep embedded clustering has become a dominating approach to unsupervised categorization of objects with deep neural networks. The optimization of the most popular methods alternates between the training of a deep autoencoder and a k-means clustering of the autoencoder{\textquoteright}s embedding. The diachronic setting, however, prevents the former to benefit from valuable information acquired by the latter. In this paper, we present an alternative where the autoencoder and the clustering are learned simultaneously. This is achieved by providing novel theoretical insight, where we show that the objective function of a certain class of Gaussian mixture models (GMM{\textquoteright}s) can naturally be rephrased as the loss function of a one-hidden layer autoencoder thus inheriting the built-in clustering capabilities of the GMM. That simple neural network, referred to as the clustering module, can be integrated into a deep autoencoder resulting in a deep clustering model able to jointly learn a clustering and an embedding. Experiments confirm the equivalence between the clustering module and Gaussian mixture models. Further evaluations affirm the empirical relevance of our deep architecture as it outperforms related baselines on several data sets.",
keywords = "Clustering, Deep autoencoders, Embedding, Gaussian mixture models, k-means, Informatics, Business informatics",
author = "Ahc{\`e}ne Boubekki and Michael Kampffmeyer and Ulf Brefeld and Robert Jenssen",
year = "2021",
month = jul,
day = "1",
doi = "10.1007/s10994-021-06015-5",
language = "English",
volume = "110",
pages = "1901--1937",
journal = "Machine Learning",
issn = "0885-6125",
publisher = "Springer Netherlands",
number = "7",

}

RIS

TY - JOUR

T1 - Joint optimization of an autoencoder for clustering and embedding

AU - Boubekki, Ahcène

AU - Kampffmeyer, Michael

AU - Brefeld, Ulf

AU - Jenssen, Robert

PY - 2021/7/1

Y1 - 2021/7/1

N2 - Deep embedded clustering has become a dominating approach to unsupervised categorization of objects with deep neural networks. The optimization of the most popular methods alternates between the training of a deep autoencoder and a k-means clustering of the autoencoder’s embedding. The diachronic setting, however, prevents the former to benefit from valuable information acquired by the latter. In this paper, we present an alternative where the autoencoder and the clustering are learned simultaneously. This is achieved by providing novel theoretical insight, where we show that the objective function of a certain class of Gaussian mixture models (GMM’s) can naturally be rephrased as the loss function of a one-hidden layer autoencoder thus inheriting the built-in clustering capabilities of the GMM. That simple neural network, referred to as the clustering module, can be integrated into a deep autoencoder resulting in a deep clustering model able to jointly learn a clustering and an embedding. Experiments confirm the equivalence between the clustering module and Gaussian mixture models. Further evaluations affirm the empirical relevance of our deep architecture as it outperforms related baselines on several data sets.

AB - Deep embedded clustering has become a dominating approach to unsupervised categorization of objects with deep neural networks. The optimization of the most popular methods alternates between the training of a deep autoencoder and a k-means clustering of the autoencoder’s embedding. The diachronic setting, however, prevents the former to benefit from valuable information acquired by the latter. In this paper, we present an alternative where the autoencoder and the clustering are learned simultaneously. This is achieved by providing novel theoretical insight, where we show that the objective function of a certain class of Gaussian mixture models (GMM’s) can naturally be rephrased as the loss function of a one-hidden layer autoencoder thus inheriting the built-in clustering capabilities of the GMM. That simple neural network, referred to as the clustering module, can be integrated into a deep autoencoder resulting in a deep clustering model able to jointly learn a clustering and an embedding. Experiments confirm the equivalence between the clustering module and Gaussian mixture models. Further evaluations affirm the empirical relevance of our deep architecture as it outperforms related baselines on several data sets.

KW - Clustering

KW - Deep autoencoders

KW - Embedding

KW - Gaussian mixture models

KW - k-means

KW - Informatics

KW - Business informatics

UR - http://www.scopus.com/inward/record.url?scp=85109174419&partnerID=8YFLogxK

U2 - 10.1007/s10994-021-06015-5

DO - 10.1007/s10994-021-06015-5

M3 - Journal articles

AN - SCOPUS:85109174419

VL - 110

SP - 1901

EP - 1937

JO - Machine Learning

JF - Machine Learning

SN - 0885-6125

IS - 7

ER -

Documents

DOI

Recently viewed

Activities

  1. Event History Analysis and Applications Using STATA - 2013
  2. Individual difference predictors of L2 learning at the syntax/semantics interface
  3. Methods and Methodologies for Researching Globalization and Development - 2015
  4. Learning to rate player actions in multi-agent scenarios
  5. Preliminary selection of experimental techniques in Subtask D
  6. The semantics of transformation: conceptual work based on Freirean methodology.
  7. Exploring Affective Human-Robot Interaction with Movie Scenes
  8. International Symposium on Multiscale Computational Analysis of Complex Materials
  9. Co-Supervisor for the Dissertation "The effects of forest structural element retention on insect communities"
  10. Co-supervisor of the dissertation "Multi-trophic interactions and functional diversity in biodiversity experiments."
  11. Methods of boundary work for inter- and transdisciplinary research.
  12. DIY Bat Detector with Gamification Elements
  13. Digitalization and cross-border knowledge transfer: The impact on international assignments
  14. Learning to Rate Player Actions on the Example of Soccer
  15. Using Gamification Elements as a tool for species protection awareness (Single Paper)
  16. Mutual Learning and Knowledge Integration in Transdisciplinary Development Teams: Empirical Findings about a Collaborative Format in Teacher Education
  17. Strengthening Form-Focused Practice in Task-Based Language Teaching Through Intelligent CALL (EUROCALL)
  18. Foresters understanding of forest nature and its gender
  19. Using the Method of Limits to Assess Comfortable Time Headways in Adaptive Cruise Control
  20. Strategies for the Control of Water Treatment by Stripping and Injection of CO2
  21. DIY-Fledermausdetektor mit Gamification-Elementen
  22. Preliminary results of the state of the art study on scaling and corrosion.
  23. Eigenzeiten of Creativity – Temporal Work as a Coordination Challenge in Artistic and Scientific Project Ecologies

Prizes

  1. Transferpreis

Publications

  1. A genetic algorithm for a self-learning parameterization of an aerodynamic part feeding system for high-speed assembly
  2. Dynamically changing sequencing rules with reinforcement learning in a job shop system with stochastic influences
  3. The Use of Factorization and Multimode Parametric Spectra in Estimating Frequency and Spectral Parameters of Signal
  4. Robust Flatness Based Control of an Electromagnetic Linear Actuator Using Adaptive PID Controller
  5. Mechanical performance prediction for friction riveting joints of dissimilar materials via machine learning
  6. Simulation based comparison of safety-stock calculation methods
  7. Modeling of lateness distributions depending on the sequencing method with respect to productivity effects
  8. A Lean Convolutional Neural Network for Vehicle Classification
  9. Editorial: Machine Learning and Data Mining in Materials Science
  10. Computing regression statistics from grouped data
  11. Understanding the error-structure of Time-driven Activity-based Costing
  12. Managing (in) times of uncertainty
  13. Optimising business performance with standard software systems
  14. Self-perception of the internal audit function within the corporate governance system - Empirical evidence for the European Union
  15. Recontextualizing Anthropomorphic Metaphors in Organization Studies
  16. Study of fuzzy controllers performance
  17. Quantification of amino acids in fermentation media by isocratic HPLC analysis of their
  18. Fusion of knowledge bases for better navigation of wheeled mobile robotic group with 3D TVS
  19. Simulation of SARS-CoV-2 pandemic in Germany with ordinary differential equations in MATLAB
  20. Validity claims in context
  21. Attention and the Speed of Information Processing
  22. Fermentative utilization of coffee mucilage using Bacillus coagulans and investigation of down-stream processing of fermentation broth for optically pure L(+)-lactic acid production
  23. Non-acceptances in context