Synthesizing Realistic Facial Expressions from Photographs

Frederic Pighin, Jamie Hecker, Dani Lischinski, David H. Salesin, and Richard Szeliski

Abstract

We present new techniques for creating photorealistic textured 3D facial models from photographs of a human subject, and for creating smooth transitions between different facial expressions by morphing between these different models. Starting from several uncalibrated views of a human subject, we employ a user-assisted technique to recover the camera poses corresponding to the views as well as the 3D coordinates of a sparse set of chosen locations on the subject’s face. A scattered data interpolation technique is then used to deform a generic face mesh to fit the particular geometry of the subject’s face. Having recovered the camera poses and the facial geometry, we extract from the input images one or more texture maps for the model. This process is repeated for several facial expressions of a particular subject. To generate transitions between these facial expressions we use 3D shape morphing between the corresponding face models, while at the same time blending the corresponding textures. Using our technique, we have been able to generate highly realistic face models and natural looking animations.

Details

Publication typeInproceedings
Published inSIGGRAPH'98
URLhttp://grail.cs.washington.edu/projects/realface/
Pages75-84
PublisherAssociation for Computing Machinery, Inc.
> Publications > Synthesizing Realistic Facial Expressions from Photographs