Clustering and Model Selection via Penalized Likelihood for Different-sized Categorical Data Vectors

Abstract : In this study, we consider unsupervised clustering of categorical vectors that can be of different size using mixture. We use likelihood maximization to estimate the parameters of the underlying mixture model and a penalization technique to select the number of mixture components. Regardless of the true distribution that generated the data, we show that an explicit penalty, known up to a multiplicative constant, leads to a non-asymptotic oracle inequality with the Kullback-Leibler divergence on the two sides of the inequality. This theoretical result is illustrated by a document clustering application. To this aim a novel robust expectation-maximization algorithm is proposed to estimate the mixture parameters that best represent the different topics. Slope heuristics are used to calibrate the penalty and to select a number of clusters.
Type de document :
Pré-publication, Document de travail
2017
Liste complète des métadonnées

Littérature citée [5 références]  Voir  Masquer  Télécharger

https://hal.inria.fr/hal-01583692
Contributeur : Erwan Le Pennec <>
Soumis le : jeudi 7 septembre 2017 - 16:51:16
Dernière modification le : jeudi 10 mai 2018 - 02:04:24

Fichiers

CategoricalVectorClustering.pd...
Fichiers produits par l'(les) auteur(s)

Identifiants

  • HAL Id : hal-01583692, version 1
  • ARXIV : 1709.02294

Citation

Esther Derman, Erwan Le Pennec. Clustering and Model Selection via Penalized Likelihood for Different-sized Categorical Data Vectors. 2017. 〈hal-01583692〉

Partager

Métriques

Consultations de la notice

307

Téléchargements de fichiers

59