“LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes” by Masso, Rogozea, Mokaram, Medvesek and Yu
Conference:
Experience Type(s):
Title:
- LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes
Organizer(s)/Presenter(s):
Description:
LIPSYNC.AI: A.I. Driven Lips and Tongue Animations Using Articulatory Phonetic Descriptors and FACS Blendshapes
References:
[1]
C. Charalambous, Zerrin Yumak, and A. Stappen. 2019. Audio‐driven emotional speech animation for interactive virtual characters. Computer Animation and Virtual Worlds 30 (2019).
Google Scholar
[2]
Daniel Cudeiro, Timo Bolkart, Cassidy Laidlaw, A. Ranjan, and Michael J. Black. 2019. Capture, Learning, and Synthesis of 3D Speaking Styles. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019), 10093–10103.
Crossref
Google Scholar
[3]
P. Edwards, Chris Landreth, M. Poplawski, R. Malinowski, Sarah Watling, E. Fiume, and Karan Singh. 2020. JALI-Driven Expressive Facial Animation and Multilingual Speech in Cyberpunk 2077. Special Interest Group on Computer Graphics and Interactive Techniques Conference Talks (2020).
Digital Library
Google Scholar
[4]
Tero Karras, Timo Aila, S. Laine, Antti Herva, and J. Lehtinen. 2017. Audio-driven facial animation by joint end-to-end learning of pose and emotion. ACM Transactions on Graphics (TOG) 36 (2017), 1 – 12.
Digital Library
Google Scholar
[5]
E. Prince, Katherine B. Martin, and D. Messinger. 2015. Facial Action Coding System.
Google Scholar
[6]
Sarah L. Taylor, Taehwan Kim, Yisong Yue, Moshe Mahler, James Krahe, Anastasio Garcia Rodriguez, J. Hodgins, and I. Matthews. 2017. A deep learning approach for generalized speech animation. ACM Transactions on Graphics (TOG) 36 (2017), 1 – 11.
Digital Library
Google Scholar
[7]
A. Thangthai, B. Milner, and Sarah L. Taylor. 2019. Synthesising visual speech using dynamic visemes and deep learning architectures. Comput. Speech Lang. 55(2019), 101–119.
Crossref
Google Scholar
[8]
Yang Zhou, Shan Xu, Chris Landreth, E. Kalogerakis, Subhransu Maji, and Karan Singh. 2018. VisemeNet: Audio-Driven Animator-Centric Speech Animation. ACM Trans. Graph. 37(2018), 161:1–161:10.
Digital Library
Google Scholar


