Yaser Sheikh
Most Recent Affiliation(s):
- Facebook Reality Labs, Carnegie Mellon University, Disney Research Pittsburgh
Other / Past Affiliation(s):
- Carnegie Mellon University
- META Corporation
Bio:
2019
Yaser Sheikh directs the Facebook Reality Lab in Pittsburgh, which is devoted to achieving photorealistic social interactions in AR and VR, and is an adjunct professor at Carnegie Mellon University. His research broadly focuses on machine perception and rendering of social behavior, spanning sub-disciplines in computer vision, computer graphics, and machine learning. With colleagues and students, he has won the Honda Initiation Award (2010), Popular Science’s “Best of What’s New” Award, best student paper award at CVPR (2018), best paper finalists at (CVPR 2019), best paper awards at WACV (2012), SAP (2012), SCA (2010), ICCV THEMIS (2009), best demo award at ECCV (2016), and he received the Hillman Fellowship for Excellence in Computer Science Research (2004). Yaser has served as a senior committee member at leading conferences in computer vision, computer graphics, and robotics including SIGGRAPH (2013, 2014), CVPR (2014, 2015, 2018), ICRA (2014, 2016), ICCP (2011), and served as an Associate Editor of CVIU. His research has been featured by various media outlets including The New York Times, BBC, MSNBC, Popular Science, and in technology media such as WIRED, The Verge, and New Scientist.
Learning Category: Jury Member:
Experience(s):
Learning Category: Presentation(s):
![Authentic volumetric avatars from a phone scan](https://history.siggraph.org/wp-content/uploads/2023/06/2022-Technical-Papers-Cao_-Authentic-Volumetric-Avatars-From-a-Phone-Scan-150x150.jpg)
Type: [Technical Papers]
Authentic volumetric avatars from a phone scan Presenter(s): [Cao] [Simon] [Kim] [Schwartz] [Zollhoefer] [Saito] [Lombardi] [Wei] [Belko] [Yu] [Sheikh] [Saragih]
[SIGGRAPH 2022]
![Frontier Talk: Metric Telepresence using Codec Avatars](https://history.siggraph.org/wp-content/uploads/2022/09/2022-SIGGRAPH-Image-Not-Available-150x150.jpg)
Type: [Frontiers]
Frontier Talk: Metric Telepresence using Codec Avatars Presenter(s): [Sheikh]
[SIGGRAPH 2022]
![State of the art in telepresence](https://history.siggraph.org/wp-content/uploads/2023/10/2022-Courses-Lawrence_State-of-the-Art-in-Telepresence-Part-1-150x150.jpg)
Type: [Courses]
State of the art in telepresence Presenter(s): [Lawrence] [Pan] [Goldman] [McDonnell] [Robillard] [O'Sullivan] [Sheikh] [Zollhoefer] [Saragih]
Entry No.: [14]
[SIGGRAPH 2022]
![Deep relightable appearance models for animatable faces](https://history.siggraph.org/wp-content/uploads/2023/06/2021-Technical-Papers-Bi_Deep-Relightable-Appearance-Models-for-Animatable-Faces-150x150.jpg)
Type: [Technical Papers]
Deep relightable appearance models for animatable faces Presenter(s): [Bi] [Lombardi] [Saito] [Simon] [Wei] [Mcphail] [Ramamoorthi] [Sheikh] [Saragih]
[SIGGRAPH 2021]
![Driving-signal aware full-body avatars](https://history.siggraph.org/wp-content/uploads/2023/06/2021-Technical-Papers-Wu_Driving-Signal-Aware-Full-Body-Avatars-150x150.jpg)
Type: [Technical Papers]
Driving-signal aware full-body avatars Presenter(s): [Bagautdinov] [Wu] [Simon] [Prada] [Shiratori] [Wei] [Xu] [Sheikh] [Saragih]
[SIGGRAPH 2021]
![Mixture of volumetric primitives for efficient neural rendering](https://history.siggraph.org/wp-content/uploads/2023/06/2021-Technical-Papers-Lombardi_Mixture-of-Volumetric-Primitives-for-Efficient-Neural-Rendering-150x150.jpg)
Type: [Technical Papers]
Mixture of volumetric primitives for efficient neural rendering Presenter(s): [Lombardi] [Simon] [Schwartz] [Zollhoefer] [Sheikh] [Saragih]
[SIGGRAPH 2021]
![Real-time 3D neural facial animation from binocular video](https://history.siggraph.org/wp-content/uploads/2023/06/2021-Technical-Papers-Cao_Real-time-3D-Neural-Facial-Animation-from-Binocular-Video-150x150.jpg)
Type: [Technical Papers]
Real-time 3D neural facial animation from binocular video Presenter(s): [Cao] [Agrawal] [Torre] [Chen] [Saragih] [Simon] [Sheikh]
[SIGGRAPH 2021]
![The Eyes Have It: An Integrated Eye and Face Model for Photorealistic Facial Animation](https://history.siggraph.org/wp-content/uploads/2022/08/2020-Technical-Papers-Schwartz_The-Eyes-Have-It-150x150.jpg)
Type: [Technical Papers]
The Eyes Have It: An Integrated Eye and Face Model for Photorealistic Facial Animation Presenter(s): [Schwartz] [Wei] [Wang] [Lombardi] [Simon] [Saragih] [Sheikh]
[SIGGRAPH 2020]
![Neural volumes: learning dynamic renderable volumes from images](https://history.siggraph.org/wp-content/uploads/2023/01/2019-Technical-Papers-Lombardi_Neural-Volumes-Learning-Dynamic-Renderable-Volumes-from-Images-150x150.jpg)
Type: [Technical Papers]
Neural volumes: learning dynamic renderable volumes from images Presenter(s): [Lombardi] [Simon] [Saragih] [Schwartz] [Lehrmann] [Sheikh]
[SIGGRAPH 2019]
![VR facial animation via multiview image translation](https://history.siggraph.org/wp-content/uploads/2023/01/2019-Technical-Papers-Wei_VR-Facial-Animation-via-Multiview-Image-Translation-150x150.jpg)
Type: [Technical Papers]
VR facial animation via multiview image translation Presenter(s): [Wei] [Saragih] [Simon] [Harley] [Lombardi] [Perdoch] [Hypes] [Wang] [Badino] [Sheikh]
[SIGGRAPH 2019]
![Deep appearance models for face rendering](https://history.siggraph.org/wp-content/uploads/2023/02/2018-Technical-Papers-Lombardi_Deep-Appearance-Models-for-Face-Rendering-150x150.jpg)
Type: [Technical Papers]
Deep appearance models for face rendering Presenter(s): [Lombardi] [Saragih] [Simon] [Sheikh]
Entry No.: [68]
[SIGGRAPH 2018]
![Gaze-Driven Video Re-Editing](https://history.siggraph.org/wp-content/uploads/2022/07/2015-SIGGRAPH-Image-Not-Available-150x150.jpg)
Type: [Technical Papers]
Gaze-Driven Video Re-Editing Presenter(s): [Jain] [Sheikh] [Shamir] [Hodgins]
[SIGGRAPH 2015]
![3D object manipulation in a single photograph using stock 3D models](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Kholgade_3D-Object-Manipulation-in-a-Single-Photograph-using-Stock-3D-Models-150x150.jpg)
Type: [Technical Papers]
3D object manipulation in a single photograph using stock 3D models Presenter(s): [Kholgade] [Simon] [Efros] [Sheikh]
[SIGGRAPH 2014]
![Automatic editing of footage from multiple social cameras](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Arev_Automatic-Editing-of-Footage-from-Multiple-Social-Cameras-150x150.jpg)
Type: [Technical Papers]
Automatic editing of footage from multiple social cameras Presenter(s): [Arev] [Park] [Sheikh] [Hodgins] [Shamir]
[SIGGRAPH 2014]
![Bilinear Spatiotemporal Basis Models](https://history.siggraph.org/wp-content/uploads/2023/03/2012-Technical-Papers-Akhter_Bilinear-Spatiotemporal-Basis-Models-150x150.jpg)
Type: [Technical Papers]
Bilinear Spatiotemporal Basis Models Presenter(s): [Akhter] [Simon] [Khan] [Matthews] [Sheikh]
[SIGGRAPH 2012]
Learning Category: Moderator:
![How do people edit light fields?](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Jarabo_How-Do-People-Edit-Light-Fields-150x150.jpg)
Type: [Technical Papers]
How do people edit light fields? Presenter(s): [Jarabo] [Masia] [Bousseau] [Pellacini] [Gutierrez]
[SIGGRAPH 2014]
![Modeling and optimizing eye vergence response to stereoscopic cuts](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Templin_Modeling-and-Optimizing-Eye-Vergence-Response-to-Stereoscopic-Cuts-150x150.jpg)
Type: [Technical Papers]
Modeling and optimizing eye vergence response to stereoscopic cuts Presenter(s): [Templin] [Didyk] [Myszkowski] [Hefeeda] [Seidel] [Matusik]
[SIGGRAPH 2014]
![Simulating and compensating changes in appearance between day and night vision](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Wanat_Simulating-and-compensating-changes-in-appearance-150x150.jpg)
Type: [Technical Papers]
Simulating and compensating changes in appearance between day and night vision Presenter(s): [Wanat] [Mantiuk]
[SIGGRAPH 2014]
![Style transfer for headshot portraits](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Shih_Style-Transfer-for-Headshot-Portraits-150x150.jpg)
Type: [Technical Papers]
Style transfer for headshot portraits Presenter(s): [Shih] [Paris] [Barnes] [Freeman] [Durand]
[SIGGRAPH 2014]
![Transient attributes for high-level understanding and editing of outdoor scenes](https://history.siggraph.org/wp-content/uploads/2023/02/2014-Technical-Papers-Laffont_Transient-Attributes-for-High-Level-Understanding-and-Editing-of-Outdoor-Scenes-150x150.jpg)
Type: [Technical Papers]
Transient attributes for high-level understanding and editing of outdoor scenes Presenter(s): [Laffont] [Ren] [Tao] [Qian] [Hays]
[SIGGRAPH 2014]
![3D shape regression for real-time facial animation](https://history.siggraph.org/wp-content/uploads/2023/03/2013-Technical-Papers-Cao_3D-Shape-Regression-for-Real-time-Facial-Animation-150x150.jpg)
Type: [Technical Papers]
3D shape regression for real-time facial animation Presenter(s): [Cao] [Weng] [Lin] [Zhou]
[SIGGRAPH 2013]
![Online modeling for realtime facial animation](https://history.siggraph.org/wp-content/uploads/2023/03/2013-Technical-Papers-Bouaziz_Online-Modeling-For-Realtime-Facial-Animation-150x150.jpg)
Type: [Technical Papers]
Online modeling for realtime facial animation Presenter(s): [Bouaziz] [Wang] [Pauly]
[SIGGRAPH 2013]
Role(s):
- Course Presenter
- Frontiers Presenter
- Studio (SIGGRAPH Lab) Presenter
- Technical Paper Moderator
- Technical Paper Presenter
- Technical Papers Jury Member