• Title/Summary/Keyword: 3차원 얼굴영상

Search Result 172, Processing Time 0.02 seconds

3D Facial Animation with Head Motion Estimation and Facial Expression Cloning (얼굴 모션 추정과 표정 복제에 의한 3차원 얼굴 애니메이션)

  • Kwon, Oh-Ryun;Chun, Jun-Chul
    • The KIPS Transactions:PartB
    • /
    • v.14B no.4
    • /
    • pp.311-320
    • /
    • 2007
  • This paper presents vision-based 3D facial expression animation technique and system which provide the robust 3D head pose estimation and real-time facial expression control. Many researches of 3D face animation have been done for the facial expression control itself rather than focusing on 3D head motion tracking. However, the head motion tracking is one of critical issues to be solved for developing realistic facial animation. In this research, we developed an integrated animation system that includes 3D head motion tracking and facial expression control at the same time. The proposed system consists of three major phases: face detection, 3D head motion tracking, and facial expression control. For face detection, with the non-parametric HT skin color model and template matching, we can detect the facial region efficiently from video frame. For 3D head motion tracking, we exploit the cylindrical head model that is projected to the initial head motion template. Given an initial reference template of the face image and the corresponding head motion, the cylindrical head model is created and the foil head motion is traced based on the optical flow method. For the facial expression cloning we utilize the feature-based method, The major facial feature points are detected by the geometry of information of the face with template matching and traced by optical flow. Since the locations of varying feature points are composed of head motion and facial expression information, the animation parameters which describe the variation of the facial features are acquired from geometrically transformed frontal head pose image. Finally, the facial expression cloning is done by two fitting process. The control points of the 3D model are varied applying the animation parameters to the face model, and the non-feature points around the control points are changed by use of Radial Basis Function(RBF). From the experiment, we can prove that the developed vision-based animation system can create realistic facial animation with robust head pose estimation and facial variation from input video image.

Pose-invariant Face Recognition using a Cylindrical Model and Stereo Camera (원통 모델과 스테레오 카메라를 이용한 포즈 변화에 강인한 얼굴인식)

  • 노진우;홍정화;고한석
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.7
    • /
    • pp.929-938
    • /
    • 2004
  • This paper proposes a pose-invariant face recognition method using cylindrical model and stereo camera. We divided this paper into two parts. One is single input image case, the other is stereo input image case. In single input image case, we normalized a face's yaw pose using cylindrical model, and in stereo input image case, we normalized a face's pitch pose using cylindrical model with previously estimated pitch pose angle by the stereo geometry. Also, since we have an advantage that we can utilize two images acquired at the same time, we can increase overall recognition performance by decision-level fusion. Through representative experiments, we achieved an increased recognition rate from 61.43% to 94.76% by the yaw pose transform, and the recognition rate with the proposed method achieves as good as that of the more complicated 3D face model. Also, by using stereo camera system we achieved an increased recognition rate 5.24% more for the case of upper face pose, and 3.34% more by decision-level fusion.

3D face recognition based on radial basis function network (방사 기저 함수 신경망을 이용한 3차원 얼굴인식)

  • Yang, Uk-Il;Sohn, Kwang-Hoon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.2 s.314
    • /
    • pp.82-92
    • /
    • 2007
  • This paper describes a novel global shape (GS) feature based on radial basis function network (RBFN) and the extraction method of the proposed feature for 3D face recognition. RBFN is the weighted sum of RBfs, it well present the non-linearity of a facial shape using the linear combination of RBFs. It is the proposed facial feature that the weights of RBFN learned by the horizontal profiles of a face. RBFN based feature expresses the locality of the facial shape even if it is GS feature, and it reduces the feature complexity like existing global methods. And it also get the smoothing effect of the facial shape. Through the experiments, we get 94.7% using the proposed feature and hidden markov model (HMM) to match the features for 100 gallery set with those for 300 test set.

Face Recognition using Contourlet Transform and PCA (Contourlet 변환 및 PCA에 의한 얼굴인식)

  • Song, Chang-Kyu;Kwon, Seok-Young;Chun, Myung-Geun
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.17 no.3
    • /
    • pp.403-409
    • /
    • 2007
  • Contourlet transform is an extention of the wavelet transform in two dimensions using the multiscale and directional fillet banks. The contourlet transform has the advantages of multiscale and time-frequency-localization properties of wavelets, but also provides a high degree of directionality. In this paper, we propose a face recognition system based on fusion methods using contourlet transform and PCA. After decomposing a face image into directional subband images by contourlet, features are obtained in each subband by PCA. Finally, face recognition is performed by fusion technique that effectively combines similarities calculated respectively In each local subband. To show the effectiveness of the proposed method, we performed experiments for ORL and CBNU dataset, and then we obtained better recognition performance in comparison with the results produced by conventional methods.

실감 통신 구현을 위한 융합 미디어 기술

  • Son, Jeong-Yeong;Lee, Hyeong;Son, Uk-Ho
    • Information and Communications Magazine
    • /
    • v.28 no.6
    • /
    • pp.39-46
    • /
    • 2011
  • 실감 통신의 구현을 위해서는 의사 소통 상대자가 서로 얼굴을 맞대고 이야기하는 것과 같은 분위기를 통신 단말을 통해 구현이 가능해야 하며, 상대방과의 상호작용도 가능해야 한다. 이러한 통신단말을 구현하기 위한 기본 조건은 연속시차에 기반한 눈에 피로감을 주지 않는 3차원 영상과 이들 영상과 접촉 및 분위기 전달이 가능한 감성 및 오감 수수 기술의 개발에 의해서만 충족된다. 이들 감성 및 오감 수수가 가능하기 위해서는 단말 자체가 디스플레이에 다수의 센서를 융합시킨 다 기능 융합미디어의 역할을 해야 하며 눈에 피로감과 시청자의 자세제한을 요구하지 않는 3차원 영상이 이 디스플레이를 통해 표시되어야한다.

Robust AAM-based Face Tracking with Occlusion Using SIFT Features (SIFT 특징을 이용하여 중첩상황에 강인한 AAM 기반 얼굴 추적)

  • Eom, Sung-Eun;Jang, Jun-Su
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.355-362
    • /
    • 2010
  • Face tracking is to estimate the motion of a non-rigid face together with a rigid head in 3D, and plays important roles in higher levels such as face/facial expression/emotion recognition. In this paper, we propose an AAM-based face tracking algorithm. AAM has been widely used to segment and track deformable objects, but there are still many difficulties. Particularly, it often tends to diverge or converge into local minima when a target object is self-occluded, partially or completely occluded. To address this problem, we utilize the scale invariant feature transform (SIFT). SIFT is an effective method for self and partial occlusion because it is able to find correspondence between feature points under partial loss. And it enables an AAM to continue to track without re-initialization in complete occlusions thanks to the good performance of global matching. We also register and use the SIFT features extracted from multi-view face images during tracking to effectively track a face across large pose changes. Our proposed algorithm is validated by comparing other algorithms under the above 3 kinds of occlusions.

Bilateral Diagonal 2DLDA Method for Human Face Recognition (얼굴 인식을 위한 쌍대각 2DLDA 방법)

  • Kim, Young-Gil;Song, Young-Jun;Kim, Dong-Woo;Ahn, Jae-Hyeong
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.19 no.5
    • /
    • pp.648-654
    • /
    • 2009
  • In this paper, a method called bilateral diagonal 2DLDA is proposed for face recognition. Two methods called Dia2DPCA and Dia2DLDA were suggested to reserve the correlations between the variations in the rows and columns of diagonal images. However, these methods work in the row direction of these images. A row-directional projection matrix can be obtained by calculating the between-class and within-class covariance matrices making an allowance for the column variation of alternative diagonal face images. In addition, column-directional projection matrix can be obtained by calculating the between-class and within-class covariance matrices making an allowance for the row variation in diagonal images. A bilateral projection scheme was applied using left and right multiplying projection matrices. As a result, the dimension of the feature matrix and computation time can be reduced. Experiments carried out on an ORL face database show that the proposed method with three different distance measures, namely, Frobenius, Yang and AMD, is more accurate than some methods, such as 2DPCA, B2DPCA, 2DLDA, etc.

Gaze Detection Using Facial Movement in Multimodal Interface (얼굴의 움직임을 이용한 다중 모드 인터페이스에서의 응시 위치 추출)

  • 박강령;남시욱;한승철;김재희
    • Proceedings of the Korean Society for Emotion and Sensibility Conference
    • /
    • 1997.11a
    • /
    • pp.168-173
    • /
    • 1997
  • 시선의 추출을 통해 사용자의 관심 방향을 알고자하는 연구는 여러 분야에 응용될 수 있는데, 대표적인 것이 장애인의 컴퓨터 이용이나, 다중 윈도우에서 마우스의 기능 대용 및, VR에서의 위치 추적 장비의 대용 그리고 원격 회의 시스템에서의 view controlling등이다. 기존의 대부분의 연구들에서는 얼굴의 입력된 동영상으로부터 얼굴의 3차원 움직임량(rotation, translation)을 구하는데 중점을 두고 있으나 [1][2], 모니터, 카메라, 얼굴 좌표계간의 복잡한 변환 과정때문에 이를 바탕으로 사용자의 응시 위치를 파악하고자하는 연구는 거으 이루어지지 않고 있다. 본 논문에서는 일반 사무실 환경에서 입력된 얼굴 동영상으로부터 얼굴 영역 및 얼굴내의 눈, 코, 입 영역 등을 추출함으로써 모니터의 일정 영역을 응시하는 순간 변화된 특징점들의 위치 및 특징점들이 형성하는 기하학적 모양의 변화를 바탕으로 응시 위치를 계산하였다. 이때 앞의 세 좌표계간의 복잡한 변환 관계를 해결하기 위하여, 신경망 구조(다층 퍼셉트론)을 이용하였다. 신경망의 학습 과정을 위해서는 모니터 화면을 15영역(가로 5등분, 세로 3등분)으로 분할하여 각 영역의 중심점을 응시할 때 추출된 특징점들을 사용하였다. 이때 학습된 15개의 응시 위치이외에 또 다른 응시 영역에 대한 출력값을 얻기 위해, 출력 함수로 연속적이고 미분가능한 함수(linear output function)를 사용하였다. 실험 결과 신경망을 이용한 응시위치 파악 결과가 선형 보간법[3]을 사용한 결과보다 정확한 성능을 나타냈다.

  • PDF

The Improvement of Meshwarp Algorithm for Rotational Pose Transformation of a Front Facial Image (정면 얼굴 영상의 회전 포즈 변형을 위한 메쉬워프 알고리즘의 개선)

  • Kim, Young-Won;Phan, Hung The;Oh, Seung-Taek;Jun, Byung-Hwan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2002.11a
    • /
    • pp.425-428
    • /
    • 2002
  • 본 논문에서는 한 장의 정면 얼굴 영상만으로 회전 변형을 수행할 수 있는 새로운 영상기반렌더링(Image Based Rendering, IBR) 기법을 제안한다. 3차원 기하학적 모델을 대신하면서 수평 회전 변형을 연출하기 위해, 특정 인물의 정면, 좌우 반측면, 좌우 측면의 얼굴 영상에 대한 표준 메쉬 집합을 작성한다. 변형하고자 하는 임의의 인물에 대해서는 정면 영상에 대한 메쉬만을 작성하고, 나머지 측면 참조 메쉬들은 표준 메쉬 집합에 의해 자동으로 생성된다. 입체적인 회전 효과를 연출하기 위해, 회전 변형시 발생할 수 있는 제어점들간의 중첩 및 역전을 허용하도록 기존의 두 단계 메쉬워프 알고리즘을 개선한 역전가능 메쉬워프 알고리즘(invertible meshwarp algorithm)을 제안한다. 이 알고리즘을 이용하여 다양한 남녀노소의 정면 얼굴 영상에 대해 회전에 따른 포즈 변형을 수행하여 비교적 자연스러운 포즈 변형 결과를 얻었다.

  • PDF

Gaze Detection System by IR-LED based Camera (적외선 조명 카메라를 이용한 시선 위치 추적 시스템)

  • 박강령
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.4C
    • /
    • pp.494-504
    • /
    • 2004
  • The researches about gaze detection have been much developed with many applications. Most previous researches only rely on image processing algorithm, so they take much processing time and have many constraints. In our work, we implement it with a computer vision system setting a IR-LED based single camera. To detect the gaze position, we locate facial features, which is effectively performed with IR-LED based camera and SVM(Support Vector Machine). When a user gazes at a position of monitor, we can compute the 3D positions of those features based on 3D rotation and translation estimation and affine transform. Finally, the gaze position by the facial movements is computed from the normal vector of the plane determined by those computed 3D positions of features. In addition, we use a trained neural network to detect the gaze position by eye's movement. As experimental results, we can obtain the facial and eye gaze position on a monitor and the gaze position accuracy between the computed positions and the real ones is about 4.2 cm of RMS error.