Predicting Tongue Positions from Acoustics and Facial Features

Asterios Toutios 1 Slim Ouni 2, 1
1 PAROLE - Analysis, perception and recognition of speech
INRIA Lorraine, LORIA - Laboratoire Lorrain de Recherche en Informatique et ses Applications
2 MULTISPEECH - Speech Modeling for Facilitating Oral-Based Communication
Inria Nancy - Grand Est, LORIA - NLPKD - Department of Natural Language Processing & Knowledge Discovery
Abstract : We test the hypothesis that adding information regarding the positions of electromagnetic articulograph (EMA) sensors on the lips and jaw can improve the results of a typical acoustic-to-EMA mapping system, based on support vector regression, that targets the tongue sensors. Our initial motivation is to use such a system in the context of adding a tongue animation to a talking head built on the basis of concatenating bimodal acoustic-visual units. For completeness, we also train a system that maps only jaw and lip information to tongue information.
Document type :
Conference papers
Complete list of metadatas

Cited literature [19 references]  Display  Hide  Download

https://hal.inria.fr/inria-00602412
Contributor : Slim Ouni <>
Submitted on : Friday, May 13, 2016 - 1:52:26 PM
Last modification on : Tuesday, December 18, 2018 - 4:38:02 PM
Long-term archiving on: Tuesday, August 16, 2016 - 9:13:02 AM

File

paper.pdf
Files produced by the author(s)

Licence


Distributed under a Creative Commons Attribution 4.0 International License

Identifiers

  • HAL Id : inria-00602412, version 1

Collections

Citation

Asterios Toutios, Slim Ouni. Predicting Tongue Positions from Acoustics and Facial Features. 12th Annual Conference of the International Speech Communication Association - Interspeech 2011, Aug 2011, Florence, Italy. ⟨inria-00602412⟩

Share

Metrics

Record views

523

Files downloads

129