Skip to Main content Skip to Navigation
Conference papers

Using multimodal speech production data to evaluate articulatory animation for audiovisual speech synthesis

Abstract : The importance of modeling speech articulation for high-quality audiovisual (AV) speech synthesis is widely acknowledged. Nevertheless, while state-of-the-art, data-driven approaches to facial animation can make use of sophisticated motion capture techniques, the animation of the intraoral articulators (viz. the tongue, jaw, and velum) typically makes use of simple rules or viseme morphing, in stark contrast to the otherwise high quality of facial modeling. Using appropriate speech production data could significantly improve the quality of articulatory animation for AV synthesis.
Complete list of metadatas

https://hal.inria.fr/hal-00734464
Contributor : Ingmar Steiner <>
Submitted on : Saturday, September 22, 2012 - 12:16:12 AM
Last modification on : Tuesday, December 18, 2018 - 4:38:02 PM
Long-term archiving on: : Friday, December 16, 2016 - 3:16:05 PM

Files

Identifiers

  • HAL Id : hal-00734464, version 1
  • ARXIV : 1209.4982

Collections

Citation

Ingmar Steiner, Korin Richmond, Slim Ouni. Using multimodal speech production data to evaluate articulatory animation for audiovisual speech synthesis. 3rd International Symposium on Facial Analysis and Animation - FAA 2012, Sep 2012, Vienna, Austria. ⟨hal-00734464⟩

Share

Metrics

Record views

407

Files downloads

332