Online Multimodal Speaker Detection for Humanoid Robots

Abstract : In this paper we address the problem of audio-visual speaker detection. We introduce an online system working on the humanoid robot NAO. The scene is perceived with two cameras and two microphones. A multimodal Gaussian mixture model (mGMM) fuses the information extracted from the auditory and visual sensors and detects the most probable audio-visual object, e.g., a person emitting a sound, in the 3D space. The system is implemented on top of a platform-independent middleware and it is able to process the information online (17Hz). A detailed description of the system and its implementation are provided, with special emphasis on the online processing issues and the proposed solutions. Experimental validation, performed with five different scenarios, show that that the proposed method opens the door to robust human-robot interaction scenarios.
Type de document :
Communication dans un congrès
Humanoids 2012 - IEEE International Conference on Humanoid Robotics, Nov 2012, Osaka, Japan. IEEE, pp.126-133, 2012, 〈10.1109/HUMANOIDS.2012.6651509〉
Liste complète des métadonnées

Littérature citée [17 références]  Voir  Masquer  Télécharger

https://hal.inria.fr/hal-00768764
Contributeur : Team Perception <>
Soumis le : dimanche 23 décembre 2012 - 19:44:32
Dernière modification le : mercredi 11 avril 2018 - 01:59:12
Document(s) archivé(s) le : dimanche 24 mars 2013 - 03:51:06

Fichier

Sanchez-Humanoids2012.pdf
Fichiers produits par l'(les) auteur(s)

Identifiants

Collections

Citation

Jordi Sanchez-Riera, Xavier Alameda-Pineda, Johannes Wienke, Antoine Deleforge, Soraya Arias, et al.. Online Multimodal Speaker Detection for Humanoid Robots. Humanoids 2012 - IEEE International Conference on Humanoid Robotics, Nov 2012, Osaka, Japan. IEEE, pp.126-133, 2012, 〈10.1109/HUMANOIDS.2012.6651509〉. 〈hal-00768764〉

Partager

Métriques

Consultations de la notice

672

Téléchargements de fichiers

308