Abstract : The problem of multimodal clustering arises whenever the data are gathered with several physically different sensors. Observations from different modalities are not necessarily aligned in the sense there there is no obvious way to associate or to compare them in some common space. A solution may consist in considering multiple clustering tasks independently for each modality. The main difficulty with such an approach is to guarantee that the unimodal clusterings are mutually consistent. In this paper we show that multimodal clustering can be addressed within a novel framework, namely conjugate mixture models. These models exploit the explicit transformations that are often available between an unobserved parameter space (objects) and each one of the observation spaces (sensors). We formulate the problem as a likelihood maximization task and we derive the associated conjugate expectation-maximization algorithm. The convergence properties of the proposed algorithm are thoroughly investigated. Several local/global optimization techniques are proposed in order to increase its convergence speed. Two initialization strategies are proposed and compared. A consistent model-selection criterion is proposed. The algorithm and its variants are tested and evaluated within the task of 3D localization of several speakers using both auditory and visual data.
Document type :
Neural Computation, Massachusetts Institute of Technology Press (MIT Press), 2011, 23 (2), pp.517-557. <10.1162/NECO_a_00074>
Contributor : Team Perception <>
Submitted on : Tuesday, May 3, 2011 - 9:53:02 AM
Last modification on : Wednesday, December 14, 2016 - 1:07:22 AM
Document(s) archivé(s) le : Thursday, August 4, 2011 - 3:09:22 AM