Paper @ Speech Communications on Automatic Animation of an Articulatory Tongue Model from Ultrasound Images of the Vocal Tract



Diandra Fabre, Thomas Hueber, Laurent Girin, Xavier Alameda-Pineda and Pierre Badin


We’ve got a paper accepted at Speech Communications on Automatic Animation of an Articulatory Tongue Model from Ultrasound Images of the Vocal Tract [1].

Abstract: Visual biofeedback is the process of gaining awareness of physiological functions through the display of visual information. As speech is concerned, visual biofeedback usually consists in showing a speaker his/her own articulatory movements, which has proven useful in applications such as speech therapy or second language learning. This article presents a novel method for automatically animating an articulatory tongue model from ultrasound images. Integrating this model into a virtual talking head enables to overcome the limitations of displaying raw ultrasound images, and provides a more complete and user-friendly feedback by showing not only the tongue, but also the palate, teeth, pharynx, etc. Altogether, these cues are expected to lead to an easier understanding of the tongue movements. Our approach is based on a probabilistic model which converts raw ultrasound images of the vocal tract into control parameters of the articulatory tongue model. We investigated several mapping techniques such as the Gaussian Mixture Regression (GMR), and in particular the Cascaded Gaussian Mixture Regression (C-GMR) techniques, recently proposed in the context of acoustic-articulatory inversion. Both techniques are evaluated on a multispeaker database. The C-GMR consists in the adaptation of a GMR reference model, trained with a large dataset of multimodal articulatory data from a reference speaker, to a new source speaker using a small set of adaptation data recorded during a preliminary enrollment session (system calibration). By using prior information from the reference model, the C-GMR approach is able (i) to maintain good mapping performance while minimizing the amount of adaptation data (and thus limiting the duration of the enrollment session), and (ii) to generalize to articulatory configurations not seen during enrollment better than the GMR approach. As a result, the C-GMR appears to be a good mapping technique for a practical system of visual biofeedback.

References:

  1. D. Fabre, T. Hueber, L. Girin, X. Alameda-Pineda, and P. Badin, “Automatic Animation of an Articulatory Tongue Model from Ultrasound Images of the Vocal Tract,” Speech Communications, 2017. [ bib pdf ]
    @article{Fabre-SpeechCom-2017,
    title={Automatic Animation of an Articulatory Tongue Model from Ultrasound Images of the Vocal Tract},
    author={Diandra Fabre and Thomas Hueber and Laurent Girin and Xavier Alameda-Pineda and Pierre Badin},
    journal={Speech Communications},
    year={2017},
      pdf={http://xavirema.eu/wp-content/papercite-data/pdf/Fabre-SpeechCom-2017.pdf}
    }

Category: Research

No responses yet.

Your email address will not be published. Required fields are marked *

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <s> <strike> <strong>