“LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes” by Masso, Rogozea, Mokaram, Medvesek and Yu – ACM SIGGRAPH HISTORY ARCHIVES

“LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes” by Masso, Rogozea, Mokaram, Medvesek and Yu

  • ©

Conference:


Experience Type(s):


Title:


    LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes

Organizer(s)/Presenter(s):



Description:


    LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes

References:


    [1]
    C. Charalambous, Zerrin Yumak, and A. Stappen. 2019. Audio‐driven emotional speech animation for interactive virtual characters. Computer Animation and Virtual Worlds 30 (2019).
    Google Scholar
    [2]
    Daniel Cudeiro, Timo Bolkart, Cassidy Laidlaw, A. Ranjan, and Michael J. Black. 2019. Capture, Learning, and Synthesis of 3D Speaking Styles. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019), 10093–10103.
    Crossref
    Google Scholar
    [3]
    P. Edwards, Chris Landreth, M. Poplawski, R. Malinowski, Sarah Watling, E. Fiume, and Karan Singh. 2020. JALI-Driven Expressive Facial Animation and Multilingual Speech in Cyberpunk 2077. Special Interest Group on Computer Graphics and Interactive Techniques Conference Talks (2020).
    Digital Library
    Google Scholar
    [4]
    Tero Karras, Timo Aila, S. Laine, Antti Herva, and J. Lehtinen. 2017. Audio-driven facial animation by joint end-to-end learning of pose and emotion. ACM Transactions on Graphics (TOG) 36 (2017), 1 – 12.
    Digital Library
    Google Scholar
    [5]
    E. Prince, Katherine B. Martin, and D. Messinger. 2015. Facial Action Coding System.
    Google Scholar
    [6]
    Sarah L. Taylor, Taehwan Kim, Yisong Yue, Moshe Mahler, James Krahe, Anastasio Garcia Rodriguez, J. Hodgins, and I. Matthews. 2017. A deep learning approach for generalized speech animation. ACM Transactions on Graphics (TOG) 36 (2017), 1 – 11.
    Digital Library
    Google Scholar
    [7]
    A. Thangthai, B. Milner, and Sarah L. Taylor. 2019. Synthesising visual speech using dynamic visemes and deep learning architectures. Comput. Speech Lang. 55(2019), 101–119.
    Crossref
    Google Scholar
    [8]
    Yang Zhou, Shan Xu, Chris Landreth, E. Kalogerakis, Subhransu Maji, and Karan Singh. 2018. VisemeNet: Audio-Driven Animator-Centric Speech Animation. ACM Trans. Graph. 37(2018), 161:1–161:10.
    Digital Library
    Google Scholar


ACM Digital Library Publication:


Overview Page:



Submit a story:

If you would like to submit a story about this experience or presentation, please contact us: historyarchives@siggraph.org