Uncertainty-based learning of Gaussian mixture models from noisy data

Abstract : We consider the problem of Gaussian mixture model (GMM)-based classification of noisy data, where the uncertainty over the data is given by a Gaussian distribution. While this uncertainty is commonly exploited at the decoding stage via uncertainty decoding, it has not been exploited at the training stage so far. We introduce a new Expectation-Maximization (EM) algorithm called uncertainty training that allows to learn GMMs directly from noisy data while taking their uncertainty into account. We evaluate its potential for a speaker recognition task over speech data corrupted by real-world domestic background noise, using a state-of-the-art signal enhancement technique and various uncertainty estimation techniques as a front-end. Compared to conventional training, the proposed algorithm results in 3\% to 4\% absolute improvement in speaker recognition accuracy by training from either matched, unmatched or multi-condition noisy data. This algorithm is also applicable with minor modifications to maximum a posteriori (MAP) or maximum likelihood linear regression (MLLR) model adaptation and to the training of hidden Markov models (HMMs) from noisy data.
Liste complète des métadonnées

https://hal.inria.fr/hal-00660689
Contributor : Emmanuel Vincent <>
Submitted on : Sunday, July 15, 2012 - 3:16:40 PM
Last modification on : Thursday, March 21, 2019 - 2:20:12 PM
Document(s) archivé(s) le : Tuesday, October 16, 2012 - 2:21:28 AM

File

RR-7862.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-00660689, version 2

Citation

Alexey Ozerov, Mathieu Lagrange, Emmanuel Vincent. Uncertainty-based learning of Gaussian mixture models from noisy data. [Research Report] RR-7862, 2012. ⟨hal-00660689v2⟩

Share

Metrics

Record views

691

Files downloads

277