On Parameterizing of Human Expression Using ICA

독립 요소 분석을 이용한 얼굴 표정의 매개변수화

  • Published : 2009.03.01

Abstract

In this paper, a novel framework that synthesizes and clones facial expression in parameter spaces is presented. To overcome the difficulties in manipulating face geometry models with high degrees of freedom, many parameterization methods have been introduced. In this paper, a data-driven parameterization method is proposed that represents a variety of expressions with a small set of fundamental independent movements based on the ICA technique. The face deformation due to the parameters is also learned from the data to capture the nonlinearity of facial movements. With this parameterization, one can control the expression of an animated character's face by the parameters. By separating the parameterization and the deformation learning process, we believe that we can adopt this framework for a variety applications including expression synthesis and cloning. The experimental result demonstrates the efficient production of realistic expressions using the proposed method.

본 논문에서는 표정을 매개변수 공간에서 표현하고 응용하기 위한 새로운 방법을 제안한다. 컴퓨터 애니메이션 분야에서 복잡한 얼굴 표정을 보다 간단하게 표현하기 위해 표정을 매개변수화 하기 위한 다양한 연구들이 이루어졌다. 본 논문에서는 사람의 얼굴 표정을 포착한 데이터를 이용하여 표정을 구성하는 간단한 몇 개의 주요 움직임을 분석한다. 이를 위해 먼저 표정 매개 변수화를 애니메이션에 효과적으로 적용하기 위해 필요한 요구사항을 정리하고, 기존 연구들의 제한점을 파악한다. 본 논문에서는 많은 양의 표정 데이터에 독립 요소 분석(independent component analysis)기법을 적응함으로써 사람의 표정을 나타내는 독립적인 움직임을 추출하고 표정의 매개변수 공간을 구성한다. 또한 얼굴의 비선형적 움직임을 보다 정확하게 근사하기 위한 변형 모델과 데이터를 기반으로 변형 모델을 학습하기 위한 방법을 제안한다. 이러한 과정을 통하여 직관적으로 일반 사용자도 쉽게 표정을 제어할수 있는 매개변수 집합과 변형 모델을 얻을 수 있다.

Keywords

References

  1. P. Ekman and W. V. Friesen, “Facial action coding system : A technique for the measurement of facial movement,” Consulting Psychologist Press: Palo Alto, 1978.
  2. L. Torresani, D. B. Yang, E. J. Alexander, and C. Bregler, “Tracking and modeling non-rigid objects with rank con straints,” in 2001 Conference on Computer Vision and Pattern Recognition (CVPR 2001), Dec. 2001, pp. 493–500.
  3. B. Choe, H. Lee, and H.-S. Ko, “Performance-driven musclebased facial animation,” The Journal of Visualization and Computer Animation, vol. 12, no. 2, pp. 67–79, May 2001.
  4. P. Joshi, W. C. Tien, M. Desbrun, and F. Pighin, “Learning controls for blend shape based realistic facial animation,” in 2003 ACM SIGGRAPH / Eurographics Symposium on Computer Animation, Aug. 2003, pp. 187–192.
  5. E. Chuang and C. Bregler, “Mood swings: expressive speech animation,” ACM Transactions on Graphics, vol. 24, no. 2, pp. 331–347, Apr. 2005.
  6. V. Blanz and T. Vetter, “A morphable model for the synthesis of 3d faces,” in Proceedings of SIGGRAPH 99, ser. Computer Graphics Proceedings, Annual Conference Series, Aug. 1999, pp. 187–194.
  7. L. Williams, “Performance driven facial animation,” in Proceedings of SIGGRAPH 90, 1990, pp. 235–242.
  8. P. Kalra, A. Mangili, N. Magnenat-Thalmann, and D. Thalmann, “Simulation of facial muscle actions based on rational free form deformations,” A. Kilgour and L. Kjelldahl, Eds., vol. 11, no. 3, 1992, pp. 59–69. https://doi.org/10.1111/1467-8659.1130059
  9. H. Li, P. Roivainen, and R. Forcheimer, “3-d motion estimation in model-based facial image coding,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 15, no. 6, June 1993, pp. 545–555.
  10. I. Essa, S. Basu, T. Darrell, and A. Pentland, “Modeling, tracking and interactive animation of faces and heads: Using input from video,” in Computer Animation ’96, June 1996, pp. 68–79.
  11. M. Escher, I. Pandzic, and N. Magnenat-Thalmann, “Facial deformations for mpeg-4,” in Computer Animation ’98, June 1998.
  12. D. Fidaleo, J.-Y. Noh, T. Kim, R. Enciso, and U.Neumann, “Classification and volume morphing for performance driven facial animation,” in International Workshop on Digital and Computational Video, 2000.
  13. L. Zhang, N. Snavely, B. Curless, and S. M. Seitz, “Spacetime faces: high resolution capture for modeling and animation,” ACM Transactions on Graphics, vol. 23, no. 3, pp. 548–558, Aug. 2004.
  14. J.-Y. Noh and U. Neumann, “Expression cloning,” in Proceedings of ACM SIGGRAPH 2001, ser. Computer Graphics Proceedings, Annual Conference Series, Aug. 2001, pp. 277– 288.
  15. D. Vlasic, M. Brand, H. Pfister, and J. Popovic, “Face transfer with multilinear models,” ACM Transactions on Graphics, vol. 24, no. 3, pp. 426–433, Aug. 2005.
  16. F. I. Parke, “Parameterized models for facial animation,” IEEE Computer Graphics & Applications, vol. 2, no. 1, pp. 61–68, Nov. 1982.
  17. Y. Lee, D. Terzopoulos, and K. Waters, “Realistic modeling for facial animation,” in Proceedings of SIGGRAPH 95, ser. Computer Graphics Proceedings, Annual Conference Series, Aug. 1995, pp. 55–62.
  18. K.Waters, “A muscle model for animating three-dimensional facial expression,” in Computer Graphics (Proceedings of SIGGRAPH 87), July 1987, pp. 17–24.
  19. K. Waters and J. Frisbie, “A coordinated muscle model for speech animation,” in Graphics Interface ’95, May 1995, pp. 163–170.
  20. S. M. Platt, “Animating facial expressions,” in Computer Graphics (Proceedings of SIGGRAPH 81), Aug. 1981, pp. 245–252.
  21. E. S. Chuang, F. Deshpande, and C. Bregler, “Facial expression space learning,” in 10th Pacific Conference on Computer Graphics and Applications, Oct. 2002, pp. 68–76.
  22. E. Sifakis, I. Neverov, and R. Fedkiw, “Automatic determination of facial muscle activations from sparse motion capture marker data,” ACM Transactions on Graphics, vol. 24, no. 3, pp. 417–425, Aug. 2005.
  23. E. Sifakis, A. Selle, A. Robinson-Mosher, and R. Fedkiw, “Simulating speech with a physics-based facial muscle model,” in 2006 ACM SIGGRAPH / Eurographics Symposium on Computer Animation, Sept. 2006, pp. 261–270.
  24. A. Hyv¨arinen and E. Oja, “Independent component analysis: Algorithms and applications,” Neural Networks, vol. 13, no. 4–5, pp. 411–430, 2000.
  25. H. Mori and J. Hoshino, “Independent component analysis of human motion,” in Proceedings of International Conference on Acoustics Speech and Signal Processing, vol. VI, 2002, pp. 3564–3567.
  26. Y. Cao, P. Faloutsos, and F. Pighin, “Unsupervised learning for speech motion editing,” in 2003 ACM SIGGRAPH / Eurographics Symposium on Computer Animation, Aug. 2003, pp. 225–231.
  27. A. Shapiro, Y. Cao, and P. Faloutsos, “Style components,” in Graphics Interface 2006, June 2006, pp. 33–39.
  28. D. Fidaleo and U. Neumann, “Analysis of co-articulation regions for performance-driven facial animation,” Computer Animation and Virtual Worlds, vol. 15, no. 1, pp. 15–26, 2004.
  29. T. Ottosson and A. Piatyszek, “Welcome to IT++.” [Online]. Available: http://itpp.sourceforge.net/
  30. G. A. Kalberer, P.Mueller, and L. V. Gool, “Speech animation using viseme space,” in Vision, Modeling, and Visualization VMV 2002. Akademische Verlagsgesellschaft Aka GmbH, Berlin, 2002, pp. 463–470.
  31. B. Yin, K.Wang, S. Ma, and J. Guo, “General face animation expression based on ica,” no. 1, 2005, pp. 181-185.