J. Westbury, P. Milenkovic, G. Weismer, and R. Kent, X-ray microbeam speech production database, The Journal of the Acoustical Society of America, vol.88, issue.S1, pp.56-56, 1990.

J. S. Perkell, M. H. Cohen, M. A. Svirsky, M. L. Matthies, I. Garabieta et al., Electromagnetic midsagittal articulometer systems for transducing speech articulatory movements, The Journal of the Acoustical Society of America, vol.92, issue.6, pp.3078-3096, 1992.

A. A. Wrench, A multi-channel/multi-speaker articulatory database for continuous speech recognition research, Phonus, 2000.

M. Stone and E. P. Davis, A head and transducer support system for making ultrasound images of tongue/jaw movement, The Journal of The Acoustical Society of America, vol.98, issue.6, pp.3107-3112, 1995.

D. H. Whalen, K. Iskarous, M. K. Tiede, D. J. Ostry, H. Lehnertlehouillier et al., The haskins optically corrected ultrasound system (hocus), 2005.

J. Dang, Estimation of vocal tract shape from speech sounds via a physiological articulatory model, Proceedings of 5th Seminar on Speech Production: Models and Data, 2000.

O. , Tongue talking: studies in intraoral speech synthesis, 2002.

G. Fant, Acoustic theory of speech production: with calculations based on X-ray studies of Russian articulations, 1970.

Z. I. Skordilis, A. Toutios, J. Töger, and S. Narayanan, Estimation of vocal tract area function from volumetric magnetic resonance imaging, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.924-928, 2017.

P. Birkholz and D. , A three-dimensional model of the vocal tract for speech synthesis, Proceedings of the 15th international congress of phonetic sciences, pp.2597-2600, 2003.

B. H. Story, Phrase-level speech simulation with an airway modulation model of speech production, Computer speech & language, vol.27, pp.989-1010, 2013.

P. Birkholz, Modeling consonant-vowel coarticulation for articulatory speech synthesis, PloS one, vol.8, issue.4, p.60603, 2013.

A. Tsukanova, B. Elie, and Y. Laprie, Articulatory speech synthesis from static context-aware articulatory targets, International Seminar on Speech Production, pp.37-47, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01937950

Y. Laprie, B. Elie, A. Tsukanova, and P. Vuissoz, Centerline articulatory models of the velum and epiglottis for articulatory synthesis of speech, 2018 26th European Signal Processing Conference (EUSIPCO), pp.2110-2114, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01921928

Y. Lim, Y. Zhu, S. G. Lingala, D. Byrd, S. Narayanan et al., 3d dynamic mri of the vocal tract during natural speech, Magnetic resonance in medicine, vol.81, issue.3, pp.1511-1520, 2019.

M. Ruthven, A. C. Freitas, R. Boubertakh, and M. E. Miquel, Application of radial grappa techniques to single-and multislice dynamic speech mri using a 16-channel neurovascular coil, Magnetic resonance in medicine, vol.82, pp.948-958, 2019.

M. Fu, M. S. Barlaz, J. L. Holtrop, J. L. Perry, D. P. Kuehn et al., High-frame-rate full-vocal-tract 3d dynamic speech imaging, Magnetic resonance in medicine, vol.77, issue.4, pp.1619-1629, 2017.

P. Mermelstein, Articulatory model for the study of speech production, The Journal of the Acoustical Society of America, vol.53, issue.4, pp.1070-1082, 1973.

S. Maeda and Y. Laprie, Vowel and prosodic factor dependent variations of vocal-tract length, InterSpeech-14th Annual Conference of the International Speech Communication Association-2013, 2013.
URL : https://hal.archives-ouvertes.fr/hal-00836829

S. Narayanan, A. Toutios, V. Ramanarayanan, A. Lammert, J. Kim et al., Real-time magnetic resonance imaging and electromagnetic articulography database for speech production research (tc), The Journal of the Acoustical Society of America, vol.136, issue.3, pp.1307-1311, 2014.

A. Toutios and S. S. Narayanan, Advances in real-time magnetic resonance imaging of the vocal tract for speech science and technology research, APSIPA Transactions on Signal and Information Processing, vol.5, 2016.

V. Ramanarayanan, S. Tilsen, M. Proctor, J. Töger, L. Goldstein et al., Analysis of speech production real-time mri, Computer Speech & Language, 2018.

M. Uecker, S. Zhang, D. Voit, A. Karaus, K. Merboldt et al., Real-time mri at a resolution of 20 ms, NMR in Biomedicine, vol.23, issue.8, pp.986-994, 2010.

A. Niebergall, S. Zhang, E. Kunay, G. Keydana, M. Job et al., Real-time mri of speaking at a resolution of 33 ms: Undersampled radial flash with nonlinear inverse reconstruction, Magnetic Resonance in Medicine, vol.69, issue.2, pp.477-485, 2013.

A. Ozerov, E. Vincent, and F. Bimbot, A general flexible framework for the handling of prior information in audio source separation, IEEE Transactions, vol.20, issue.4, pp.1118-1133, 2012.
URL : https://hal.archives-ouvertes.fr/hal-00626962

S. Roekhaut, S. Brognaux, R. Beaufort, and T. Dutoit, eLite-HTS: Un outil TAL pour la génération de synthèse hmm en français, 2014.

S. Young, G. Evermann, M. Gales, T. Hain, D. Kershaw et al., The HTK book, Cambridge university engineering department, vol.3, p.175, 2002.

Z. Wu, O. Watts, and S. King, Merlin: An open source neural network speech synthesis system, Proc. SSW, 2016.

J. Thirion, Image matching as a diffusion process: an analogy with maxwell's demons, Medical image analysis, vol.2, issue.3, pp.243-260, 1998.

T. Vercauteren, X. Pennec, A. Perchant, and N. Ayache, Diffeomorphic demons: Efficient non-parametric image registration, NeuroImage, vol.45, issue.1, pp.61-72, 2009.
URL : https://hal.archives-ouvertes.fr/inserm-00349600

Z. Fagyal, D. Kibbee, and F. Jenkins, French: A linguistic introduction, 2006.