On-the-fly audio source separation

Abstract : This paper addresses the challenging task of single channel audio source separation. We introduce a novel concept of on-the-fly audio source separation which greatly simplifies the user's interaction with the system compared to the state-of-the-art user-guided approaches. In the proposed framework, the user is only asked to listen to an audio mixture and type some keywords (e.g. "dog barking", "wind", etc.) describing the sound sources to be separated. These keywords are then used as text queries to search for audio examples from the internet to guide the separation process. In particular, we propose several approaches to efficiently exploit these retrieved examples, including an approach based on a generic spectral model with group sparsity-inducing constraints. Finally, we demonstrate the effectiveness of the proposed framework with mixtures containing various types of sounds.
Liste complète des métadonnées

Cited literature [19 references]  Display  Hide  Download

https://hal.inria.fr/hal-01023221
Contributor : Alexey Ozerov <>
Submitted on : Friday, July 11, 2014 - 4:10:54 PM
Last modification on : Monday, July 14, 2014 - 8:52:53 AM
Document(s) archivé(s) le : Saturday, October 11, 2014 - 1:05:10 PM

File

ElBadawy_et_al_2014.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-01023221, version 1

Citation

Dalia El Badawy, Ngoc Duong, Alexey Ozerov. On-the-fly audio source separation. the 24th IEEE International Workshop on Machine Learning for Signal Processing (MLSP 2014), Sep 2014, Reims, France. ⟨hal-01023221⟩

Share

Metrics

Record views

368

Files downloads

953