“Simultaneous speech and animation synthesis” by Schabus, Pucher and Hofer

  • ©Dietmar Schabus, Michael Pucher, and Gregor Hofer




    Simultaneous speech and animation synthesis



    Talking computer animated characters are a common sight in video games and movies. Although doing the mouth animation by hand gives the best results, because of cost and time constraints it is not always feasible. Furthermore the amount of speech in current games is ever increasing with some games having more than 200,000 lines of dialogue. This work proposes a system that can produce speech and the corresponding lip animation simultaneously using a statistical machine learning framework based on Hidden Markov Models (HMMs). The key point is that with the developed system never before seen or heard animated dialogues can be produced at a push of a button.


    1. Tokuda, K., Yoshimura, T., Masuko, T., Kobayashi, T., and Kitamura, T. 2000. Speech parameter generation algorithms for HMM-based speech synthesis. In Proc. ICASSP, 1315–1318.

ACM Digital Library Publication:

Overview Page: