• Title/Summary/Keyword: Active Appearance Model

Search Result 69, Processing Time 0.031 seconds

Markerless Image-to-Patient Registration Using Stereo Vision : Comparison of Registration Accuracy by Feature Selection Method and Location of Stereo Bision System (스테레오 비전을 이용한 마커리스 정합 : 특징점 추출 방법과 스테레오 비전의 위치에 따른 정합 정확도 평가)

  • Joo, Subin;Mun, Joung-Hwan;Shin, Ki-Young
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.53 no.1
    • /
    • pp.118-125
    • /
    • 2016
  • This study evaluates the performance of image to patient registration algorithm by using stereo vision and CT image for facial region surgical navigation. For the process of image to patient registration, feature extraction and 3D coordinate calculation are conducted, and then 3D CT image to 3D coordinate registration is conducted. Of the five combinations that can be generated by using three facial feature extraction methods and three registration methods on stereo vision image, this study evaluates the one with the highest registration accuracy. In addition, image to patient registration accuracy was compared by changing the facial rotation angle. As a result of the experiment, it turned out that when the facial rotation angle is within 20 degrees, registration using Active Appearance Model and Pseudo Inverse Matching has the highest accuracy, and when the facial rotation angle is over 20 degrees, registration using Speeded Up Robust Features and Iterative Closest Point has the highest accuracy. These results indicate that, Active Appearance Model and Pseudo Inverse Matching methods should be used in order to reduce registration error when the facial rotation angle is within 20 degrees, and Speeded Up Robust Features and Iterative Closest Point methods should be used when the facial rotation angle is over 20 degrees.

3D FACE RECONSTRUCTION FROM ROTATIONAL MOTION

  • Sugaya, Yoshiko;Ando, Shingo;Suzuki, Akira;Koike, Hideki
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.714-718
    • /
    • 2009
  • 3D reconstruction of a human face from an image sequence remains an important problem in computer vision. We propose a method, based on a factorization algorithm, that reconstructs a 3D face model from short image sequences exhibiting rotational motion. Factorization algorithms can recover structure and motion simultaneously from one image sequence, but they usually require that all feature points be well tracked. Under rotational motion, however, feature tracking often fails due to occlusion and frame out of features. Additionally, the paucity of images may make feature tracking more difficult or decrease reconstruction accuracy. The proposed 3D reconstruction approach can handle short image sequences exhibiting rotational motion wherein feature points are likely to be missing. We implement the proposal as a reconstruction method; it employs image sequence division and a feature tracking method that uses Active Appearance Models to avoid the failure of feature tracking. Experiments conducted on an image sequence of a human face demonstrate the effectiveness of the proposed method.

  • PDF

Development of Facial Expression Recognition System based on Bayesian Network using FACS and AAM (FACS와 AAM을 이용한 Bayesian Network 기반 얼굴 표정 인식 시스템 개발)

  • Ko, Kwang-Eun;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.19 no.4
    • /
    • pp.562-567
    • /
    • 2009
  • As a key mechanism of the human emotion interaction, Facial Expression is a powerful tools in HRI(Human Robot Interface) such as Human Computer Interface. By using a facial expression, we can bring out various reaction correspond to emotional state of user in HCI(Human Computer Interaction). Also it can infer that suitable services to supply user from service agents such as intelligent robot. In this article, We addresses the issue of expressive face modeling using an advanced active appearance model for facial emotion recognition. We consider the six universal emotional categories that are defined by Ekman. In human face, emotions are most widely represented with eyes and mouth expression. If we want to recognize the human's emotion from this facial image, we need to extract feature points such as Action Unit(AU) of Ekman. Active Appearance Model (AAM) is one of the commonly used methods for facial feature extraction and it can be applied to construct AU. Regarding the traditional AAM depends on the setting of the initial parameters of the model and this paper introduces a facial emotion recognizing method based on which is combined Advanced AAM with Bayesian Network. Firstly, we obtain the reconstructive parameters of the new gray-scale image by sample-based learning and use them to reconstruct the shape and texture of the new image and calculate the initial parameters of the AAM by the reconstructed facial model. Then reduce the distance error between the model and the target contour by adjusting the parameters of the model. Finally get the model which is matched with the facial feature outline after several iterations and use them to recognize the facial emotion by using Bayesian Network.

Automatic Generation of Rule-based Caricature Image (규칙 기반 캐리커쳐 자동 생성 기법)

  • Lee, Eun-Jung;Kwon, Ji-Yong;Lee, In-Kwon
    • Journal of the Korea Computer Graphics Society
    • /
    • v.12 no.4
    • /
    • pp.17-22
    • /
    • 2006
  • We present the technique that automatically generates caricatures from input face images. We get the mean-shape of training images and extract input image's feature point using AAM(Active Appearance Model). From literature of caricature artists, we define exaggeration rules and apply our rules to input feature points, then we can get exaggerated feature points. To change our results into cartoon-like images, we apply some cartoon-stylizing method to input image and combine it with facial sketch. The input image is warped to exaggerated feature point for final results. Our method can automatically generate a caricature image while it minimizes user interaction.

  • PDF

Study of Emotion Recognition based on Facial Image for Emotional Rehabilitation Biofeedback (정서재활 바이오피드백을 위한 얼굴 영상 기반 정서인식 연구)

  • Ko, Kwang-Eun;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.10
    • /
    • pp.957-962
    • /
    • 2010
  • If we want to recognize the human's emotion via the facial image, first of all, we need to extract the emotional features from the facial image by using a feature extraction algorithm. And we need to classify the emotional status by using pattern classification method. The AAM (Active Appearance Model) is a well-known method that can represent a non-rigid object, such as face, facial expression. The Bayesian Network is a probability based classifier that can represent the probabilistic relationships between a set of facial features. In this paper, our approach to facial feature extraction lies in the proposed feature extraction method based on combining AAM with FACS (Facial Action Coding System) for automatically modeling and extracting the facial emotional features. To recognize the facial emotion, we use the DBNs (Dynamic Bayesian Networks) for modeling and understanding the temporal phases of facial expressions in image sequences. The result of emotion recognition can be used to rehabilitate based on biofeedback for emotional disabled.

Face Alignment using Template Warping BAM (템플릿 워핑 BAM을 이용한 얼굴 윤곽선 검출)

  • Kim, Seok-Ho;Kim, Jae-Min;Cho, Seoung-Won;Lee, Ki-Sung;Chung, Sun-Tae
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2008.04a
    • /
    • pp.418-420
    • /
    • 2008
  • 얼굴 윤곽선 검출을 위해 그동안 많은 알고리즘이 연구되었다. 그리고 최근에 기존 Active Appearance Model(AAM)에 비해 성능이 개선된 Boosted Appearance Model (BAM)가 Liu에 의해제안되었다. BAM에서는 매 반복 단계마다 Steepest Descent 영상을 구해야 하는데 입력영상의 워핑을 해야 하므로 이것은 계산량이 많다. 본 논문은 BAM을 사용하면서 매번 계산되어야 하는 입력 영상의 워핑을 대신해 템플릿이 워핑함으로써 계산 시간을 줄일 수 있는 방법을 제시한다. 템플릿은 약한 분류기에 사용되는 Haar-like feature들로 이것은 입력 영상에 비해 크기가 매우 작으므로 제안된 방법을 사용하면 Steepest Descent 영상을 구하는데 필요한 워핑 속도를 줄일 수 있다.

  • PDF

A Study on Facial Wrinkle Detection using Active Appearance Models (AAM을 이용한 얼굴 주름 검출에 관한 연구)

  • Lee, Sang-Bum;Kim, Tae-Mook
    • Journal of Digital Convergence
    • /
    • v.12 no.7
    • /
    • pp.239-245
    • /
    • 2014
  • In this paper, a weighted value wrinkle detection method is suggested based on the analysis on the entire facial features such as face contour, face size, eyes and ears. Firstly, the main facial elements are detected with AAM method entirely from the input screen images. Such elements are mainly composed of shape-based and appearance methods. These are used for learning the facial model and for matching the face from new screen images based on the learned models. Secondly, the face and background are separated in the screen image. Four points with the biggest possibilities for wrinkling are selected from the face and high wrinkle weighted values are assigned to them. Finally, the wrinkles are detected by applying Canny edge algorithm for the interested points of weighted value. The suggested algorithm adopts various screen images for experiment. The experiments display the excellent results of face and wrinkle detection in the most of the screen images.

A Study on Enhancing the Performance of Detecting Lip Feature Points for Facial Expression Recognition Based on AAM (AAM 기반 얼굴 표정 인식을 위한 입술 특징점 검출 성능 향상 연구)

  • Han, Eun-Jung;Kang, Byung-Jun;Park, Kang-Ryoung
    • The KIPS Transactions:PartB
    • /
    • v.16B no.4
    • /
    • pp.299-308
    • /
    • 2009
  • AAM(Active Appearance Model) is an algorithm to extract face feature points with statistical models of shape and texture information based on PCA(Principal Component Analysis). This method is widely used for face recognition, face modeling and expression recognition. However, the detection performance of AAM algorithm is sensitive to initial value and the AAM method has the problem that detection error is increased when an input image is quite different from training data. Especially, the algorithm shows high accuracy in case of closed lips but the detection error is increased in case of opened lips and deformed lips according to the facial expression of user. To solve these problems, we propose the improved AAM algorithm using lip feature points which is extracted based on a new lip detection algorithm. In this paper, we select a searching region based on the face feature points which are detected by AAM algorithm. And lip corner points are extracted by using Canny edge detection and histogram projection method in the selected searching region. Then, lip region is accurately detected by combining color and edge information of lip in the searching region which is adjusted based on the position of the detected lip corners. Based on that, the accuracy and processing speed of lip detection are improved. Experimental results showed that the RMS(Root Mean Square) error of the proposed method was reduced as much as 4.21 pixels compared to that only using AAM algorithm.

Three-Dimensional Active Shape Models for Medical Image Segmentation (의료영상 분할을 위한 3차원 능동 모양 모델)

  • Lim, Seong-Jae;Jeong, Yong-Yeon;Ho, Yo-Sung
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.44 no.5
    • /
    • pp.55-61
    • /
    • 2007
  • In this paper, we propose a three-dimensional(3D) active shape models for medical image segmentation. In order to build a 3D shape model, we need to generate a point distribution model(PDM) and select corresponding landmarks in all the training shapes. The manual determination method, two-dimensional(2D) method, and limited 3D method of landmark correspondences are time-consuming, tedious, and error-prone. In this paper, we generate a 3D statistical shape model using the 3D model generation method of a distance transform and a tetrahedron method for landmarking. After generating the 3D model, we extend the shape model training and gray-level model training of 2D active shape models(ASMs) and we use the integrated modeling process with scale and gray-level models for the appearance profile to represent the local structure. Experimental results are comparable to those of region-based, contour-based methods, and 2D ASMs.

A 3D Face Generation Method using Single Frontal Face Image for Game Users (단일 정면 얼굴 영상을 이용한 게임 사용자의 3차원 얼굴 생성 방법)

  • Jeong, Min-Yi;Lee, Sung-Joo;Park, Kang-Ryong;Kim, Jai-Hie
    • Proceedings of the IEEK Conference
    • /
    • 2008.06a
    • /
    • pp.1013-1014
    • /
    • 2008
  • In this paper, we propose a new method of generating 3D face by using single frontal face image and 3D generic face model. By using active appearance model (AAM), the control points among facial feature points were localized in the 2D input face image. Then, the transform parameters of 3D generic face model were found to minimize the error between the 2D control points and the corresponding 2D points projected from 3D facial model. Finally, by using the obtained model parameters, 3D face was generated. We applied this 3D face to 3D game framework and found that the proposed method could make a realistic 3D face of game user.

  • PDF