Speech driven realistic mouth animation based on multi-modal unit selection

Dongmei Jiang, Ilse Ravyse, Hichem Sahli, Werner Verhelst

Research output: Contribution to journalArticle

7 Citations (Scopus)


This paper presents a novel audio visual diviseme (viseme pair) instance selection and concatenation method for speech driven photo realistic mouth animation. Firstly, an audio visual diviseme database is built consisting of the audio feature sequences, intensity sequences and visual feature sequences of the instances. In the Viterbi based diviseme instance selection, we set the accumulative cost as the weighted sum of three items: 1) logarithm of concatenation smoothness of the synthesized mouth trajectory; 2) logarithm of the pronunciation distance; 3) logarithm of the audio intensity distance between the candidate diviseme instance and the target diviseme segment in the incoming speech. The selected diviseme instances are time warped and blended to construct the mouth animation. Objective and subjective evaluations on the synthesized mouth animations prove that the multimodal diviseme instance selection algorithm proposed in this paper outperforms the triphone unit selection algorithm in Video Rewrite. Clear, accurate, smooth mouth animations can be obtained matching well with the pronunciation and intensity changes in the incoming speech. Moreover, with the logarithm function in the accumulative cost, it is easy to set the weights to obtain optimal mouth animations.
Original languageEnglish
Article number157
Number of pages13
JournalJournal on Multimodal User Interfaces
Publication statusPublished - 2009


  • Mouth animation
  • Audio visual diviseme instance selection
  • Concatenation smoothness
  • Pronunciation distance
  • Intensity distance

Fingerprint Dive into the research topics of 'Speech driven realistic mouth animation based on multi-modal unit selection'. Together they form a unique fingerprint.

Cite this