Expressive Expression Mapping With Ratio Images
Expressive Expression Mapping With Ratio Images
net/publication/220720997
CITATIONS READS
225 230
3 authors:
Zhengyou Zhang
Microsoft
371 PUBLICATIONS 36,243 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Zhengyou Zhang on 04 January 2016.
Microsoft Research∗
1 Introduction
Facial expressions exhibit not only facial feature motions,
but also subtle changes in illumination and appearance (e.g., Figure 1: Expression mapping with expression details. Left:
facial creases and wrinkles). These details are important neutral face. Middle: result from geometric warping. Right:
visual cues, but they are difficult to synthesize. result from our method.
One class of methods to generate facial expressions with
details is the morph-based approaches and their exten-
sions [2, 14, 16, 3]. The main limitation is that this ap-
proach can only generate expressions in-between the given As an example, Figure 1 shows a comparison of an expres-
expressions through interpolation. If we only have someone’s sion with and without the expression details. The left image
neutral face, we would not be able to generate this person’s is the original neutral face. The one in the middle is the ex-
facial expressions using morph-based methods. pression generated using the traditional expression mapping
Another popular class of techniques, known as expression method. The image on the right is the expression generated
mapping ( performance-driven animation) [4, 10, 20, 13], using the method to be described in this paper. The feature
does not have such limitation. It can be used to animate 2D locations on the right image are exactly the same as those
drawings and images, as well as textured or non-textured 3D on the middle image, but because there are expression de-
face models. The method is very simple. Given an image of a tails, the right image looks much more convincing than the
person’s neutral face and another image of the same person’s middle one.
face with an expression. The positions of the face features
In this paper, we present a novel technique to capture the
(eyes, eye brows, mouths, etc.) on both images are located
illumination change of one person’s expression and map it to
either manually or through some automatic method. The
any different person to generate more expressive facial ex-
difference vector is then added to a new face’s feature posi-
pressions. The critical observation is that the illumination
tions to generate the new expression for that face through
change resulting from the change of surface normal can be
geometry-controlled image warping [21, 2, 10]. One problem
extracted in a skin-color independent manner by using what
∗ [email protected],[email protected],[email protected] we call an expression ratio image (ERI). This ERI can then
be applied to any different person to generate correct illumi-
nation changes resulted from the geometric deformation of
that person’s face.
Figure 13: Expression mapping may fail when the poses are
too far apart. (a), (b), and (c): neutral faces with different
poses. (e): result of mapping (d) to (b). (f): result of
mapping (d) to (c).
7 Limitations and future directions [12] S. R. Marschner, B. Guenter, and S. Raghupathy. Modeling
and rendering for realistic facial animation. In Rendering Tech-
One limitation of this method is in dealing with different niques, pages 231–242. Springer Wien New York, 2000.
lighting conditions. Even though we had some success of [13] F. I. Parke and K. Waters. Computer Facial Animation.
applying ERI to expressions under different lighting envi- AKPeters, Wellesley, Massachusetts, 1996.
ronment with the help of histogram matching, a more gen- [14] F. Pighin, J. Hecker, D. Lischinski, R. Szeliski, and D. H. Salesin.
eral solution is to use advanced relighting techniques such Synthesizing realistic facial expressions from photographs. In
as [6, 12, 15]. Computer Graphics, Annual Conference Series, pages 75–84.
Currently the image marks are very sparse. It is desir- Siggraph, July 1998.
able to add line and curve features for better facial feature [15] T. Riklin-Raviv and A. Shashua. The quotient image: Class
correspondences. We are planning to implement better im- based re-rendering and recongnition with varying illuminations.
age warping techniques such as those reported in [2, 10]. In IEEE Conference on Computer Vision and Pattern Recog-
Better image warping should reduce the artifacts of the nition, pages 566–571, June 1999.
triangulation-based warping method that we currently use. [16] S. M. Seize and C. R. Dyer. View morphing. In Computer
We also hope that better image warping technique together Graphics, pages 21–30. Siggraph, August 1996.
with line and curve features will improve pixel correspon-
[17] A. Stoschek. Image-based re-rendering of faces for continuous
dences. High-quality pixel correspondences could reduce the pose and illumination directions. In IEEE Conference on Com-
need of ratio image filtering, thus allowing more expression puter Vision and Pattern Recognition, pages 582–587, 2000.
details to be captured in an ERI.
[18] D. Terzopoulos and K. Waters. Physically-based facial modeling
and animation. Journal of Visualization and Computer Ani-
Acknowledgements mation, 1(4):73–80, March 1990.
[19] K. Waters. A muscle model for animating three-dimensional fa-
We would like to thank Ko Nishino for helpful discussions cial expression. Computer Graphics, 22(4):17–24, 1987.
and many helps with image acquisition. We would like to [20] L. Williams. Performace-driven facial animation. In Computer
thank Conal Elliott for carefully reading our manuscripts Graphics, pages 235–242. Siggraph, August 1990.
and providing valuable comments. We would like to thank
[21] G. Wolberg. Digital Image Warping. IEEE Computer Society
Alex Colburn, Steve Harris, Chuck Jacobs, Brian Meyer, Press, 1990.
Sing Bing Kang, Sashi Raghupathy, and Emiko Unno for
their help with image acquisition. We would like to thank
Michael Cohen for his support.
References
[1] N. Badler and S. Platt. Animating facial expressions. In Com-
puter Graphics, pages 245–252. Siggraph, August 1981.
[2] T. Beier and S. Neely. Feature-based image metamorphosis. In
Computer Graphics, pages 35–42. Siggraph, July 1992.
[3] C. Bregler, M. Covell, and M. Slaney. Video rewrite: Driving
visual speech with audio. In Computer Graphics, pages 353–
360. Siggraph, August 1997.
[4] S. E. Brennan. Caricature Generator. M.S. Visual Studies, Dept
of Architecture, Massachusetts Institute of Technology, Cam-
bridge, MA., Sept. 1982.
[5] P. E. Debevec. Rendering synthetic objects into real scenes:
Bridging traditional and image-based graphics with global illu-
mination and high dynamic range photography. In Computer
Graphics, Annual Conference Series, pages 189–198. Siggraph,
July 1998.
[6] P. E. Debevec, T. Hawkins, C. Tchou, H.-P. Duiker, W. Sarokin,
and M. Sagar. Acquiring the reflectance field of a human face. In
Computer Graphics, Annual Conference Series, pages 145–156.
Siggraph, July 2000.