• Title/Summary/Keyword: facial region detection

Search Result 116, Processing Time 0.027 seconds

Face Detection Using Adaboost and Template Matching of Depth Map based Block Rank Patterns (Adaboost와 깊이 맵 기반의 블록 순위 패턴의 템플릿 매칭을 이용한 얼굴검출)

  • Kim, Young-Gon;Park, Rae-Hong;Mun, Seong-Su
    • Journal of Broadcast Engineering
    • /
    • v.17 no.3
    • /
    • pp.437-446
    • /
    • 2012
  • A face detection algorithms using two-dimensional (2-D) intensity or color images have been studied for decades. Recently, with the development of low-cost range sensor, three-dimensional (3-D) information (i.e., depth image that represents the distance between a camera and objects) can be easily used to reliably extract facial features. Most people have a similar pattern of 3-D facial structure. This paper proposes a face detection method using intensity and depth images. At first, adaboost algorithm using intensity image classifies face and nonface candidate regions. Each candidate region is divided into $5{\times}5$ blocks and depth values are averaged in each block. Then, $5{\times}5$ block rank pattern is constructed by sorting block averages of depth values. Finally, candidate regions are classified as face and nonface regions by matching the constructed depth map based block rank patterns and a template pattern that is generated from training data set. For template matching, the $5{\times}5$ template block rank pattern is prior constructed by averaging block ranks using training data set. The proposed algorithm is tested on real images obtained by Kinect range sensor. Experimental results show that the proposed algorithm effectively eliminates most false positives with true positives well preserved.

Detection of Faces with Partial Occlusions using Statistical Face Model (통계적 얼굴 모델을 이용한 부분적으로 가려진 얼굴 검출)

  • Seo, Jeongin;Park, Hyeyoung
    • Journal of KIISE
    • /
    • v.41 no.11
    • /
    • pp.921-926
    • /
    • 2014
  • Face detection refers to the process extracting facial regions in an input image, which can improve speed and accuracy of recognition or authorization system, and has diverse applicability. Since conventional works have tried to detect faces based on the whole shape of faces, its detection performance can be degraded by occlusion made with accessories or parts of body. In this paper we propose a method combining local feature descriptors and probability modeling in order to detect partially occluded face effectively. In training stage, we represent an image as a set of local feature descriptors and estimate a statistical model for normal faces. When the test image is given, we find a region that is most similar to face using our face model constructed in training stage. According to experimental results with benchmark data set, we confirmed the effect of proposed method on detecting partially occluded face.

Automatic Face Region Detection and Tracking for Robustness in Rotation using the Estimation Function (평가 함수를 사용하여 회전에 강건한 자동 얼굴 영역 검출과 추적)

  • Kim, Ki-Sang;Kim, Gye-Young;Choi, Hyung-Il
    • The Journal of the Korea Contents Association
    • /
    • v.8 no.9
    • /
    • pp.1-9
    • /
    • 2008
  • In this paper, we proposed automatic face detection and tracking which is robustness in rotation. To detect a face image in complicated background and various illuminating conditions, we used face skin color detection. we used Harris corner detector for extract facial feature points. After that, we need to track these feature points. In traditional method, Lucas-Kanade feature tracker doesn't delete useless feature points by occlusion in current scene (face rotation or out of camera). So we proposed the estimation function, which delete useless feature points. The method of delete useless feature points is estimation value at each pyramidal level. When the face was occlusion, we deleted these feature points. This can be robustness to face rotation and out of camera. In experimental results, we assess that using estimation function is better than traditional feature tracker.

Study on Remote Face Recognition System Using by Multi Thread on Distributed Processing Server (분산처리서버에서의 멀티 쓰레드 방식을 적용한 원격얼굴인식 시스템)

  • Kim, Eui-Sun;Ko, Il-Ju
    • The Journal of Korean Institute of Next Generation Computing
    • /
    • v.13 no.5
    • /
    • pp.19-28
    • /
    • 2017
  • Various methods for reducing the load on the server have been implemented in performing face recognition remotely by the spread of IP security cameras. In this paper, IP surveillance cameras at remote sites are input through a DSP board equipped with face detection function, and then face detection is performed. Then, the facial region image is transmitted to the server, and the face recognition processing is performed through face recognition distributed processing. As a result, the overall server system load and significantly reduce processing and real-time face recognition has the advantage that you can perform while linked up to 256 cameras. The technology that can accomplish this is to perform 64-channel face recognition per server using distributed processing server technology and to process face search results through 250 camera channels when operating four distributed processing servers there was.

3D Face Alignment and Normalization Based on Feature Detection Using Active Shape Models : Quantitative Analysis on Aligning Process (ASMs을 이용한 특징점 추출에 기반한 3D 얼굴데이터의 정렬 및 정규화 : 정렬 과정에 대한 정량적 분석)

  • Shin, Dong-Won;Park, Sang-Jun;Ko, Jae-Pil
    • Korean Journal of Computational Design and Engineering
    • /
    • v.13 no.6
    • /
    • pp.403-411
    • /
    • 2008
  • The alignment of facial images is crucial for 2D face recognition. This is the same to facial meshes for 3D face recognition. Most of the 3D face recognition methods refer to 3D alignment but do not describe their approaches in details. In this paper, we focus on describing an automatic 3D alignment in viewpoint of quantitative analysis. This paper presents a framework of 3D face alignment and normalization based on feature points obtained by Active Shape Models (ASMs). The positions of eyes and mouth can give possibility of aligning the 3D face exactly in three-dimension space. The rotational transform on each axis is defined with respect to the reference position. In aligning process, the rotational transform converts an input 3D faces with large pose variations to the reference frontal view. The part of face is flopped from the aligned face using the sphere region centered at the nose tip of 3D face. The cropped face is shifted and brought into the frame with specified size for normalizing. Subsequently, the interpolation is carried to the face for sampling at equal interval and filling holes. The color interpolation is also carried at the same interval. The outputs are normalized 2D and 3D face which can be used for face recognition. Finally, we carry two sets of experiments to measure aligning errors and evaluate the performance of suggested process.

Rotation and Scale Invariant Face Detection Using Log-polar Mapping and Face Features (Log-polar변환과 얼굴특징추출을 이용한 크기 및 회전불변 얼굴인식)

  • Go Gi-Young;Kim Doo-Young
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.6 no.1
    • /
    • pp.15-22
    • /
    • 2005
  • In this paper, we propose a face recognition system by using the CCD color image. We first get the face candidate image by using YCbCr color model and adaptive skin color information. And we use it initial curve of active contour model to extract face region. We use the Eye map and mouth map using color information for extracting facial feature from the face image. To obtain center point of Log-polar image, we use extracted facial feature from the face image. In order to obtain feature vectors, we use extracted coefficients from DCT and wavelet transform. To show the validity of the proposed method, we performed a face recognition using neural network with BP learning algorithm. Experimental results show that the proposed method is robuster with higher recogntion rate than the conventional method for the rotation and scale variant.

  • PDF

A Study on Fast Iris Detection for Iris Recognition in Mobile Phone (휴대폰에서의 홍채인식을 위한 고속 홍채검출에 관한 연구)

  • Park Hyun-Ae;Park Kang-Ryoung
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.2 s.308
    • /
    • pp.19-29
    • /
    • 2006
  • As the security of personal information is becoming more important in mobile phones, we are starting to apply iris recognition technology to these devices. In conventional iris recognition, magnified iris images are required. For that, it has been necessary to use large magnified zoom & focus lens camera to capture images, but due to the requirement about low size and cost of mobile phones, the zoom & focus lens are difficult to be used. However, with rapid developments and multimedia convergence trends in mobile phones, more and more companies have built mega-pixel cameras into their mobile phones. These devices make it possible to capture a magnified iris image without zoom & focus lens. Although facial images are captured far away from the user using a mega-pixel camera, the captured iris region possesses sufficient pixel information for iris recognition. However, in this case, the eye region should be detected for accurate iris recognition in facial images. So, we propose a new fast iris detection method, which is appropriate for mobile phones based on corneal specular reflection. To detect specular reflection robustly, we propose the theoretical background of estimating the size and brightness of specular reflection based on eye, camera and illuminator models. In addition, we use the successive On/Off scheme of the illuminator to detect the optical/motion blurring and sunlight effect on input image. Experimental results show that total processing time(detecting iris region) is on average 65ms on a Samsung SCH-S2300 (with 150MHz ARM 9 CPU) mobile phone. The rate of correct iris detection is 99% (about indoor images) and 98.5% (about outdoor images).

An Automatic Strabismus Screening Method with Corneal Light Reflex based on Image Processing

  • Huang, Xi-Lang;Kim, Chang Zoo;Choi, Seon Han
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.5
    • /
    • pp.642-650
    • /
    • 2021
  • Strabismus is one of the most common disease that might be associated with vision impairment. Especially in infants and children, it is critical to detect strabismus at an early age because uncorrected strabismus may go on to develop amblyopia. To this end, ophthalmologists usually perform the Hirschberg test, which observes corneal light reflex (CLR) to determine the presence and type of strabismus. However, this test is usually done manually in a hospital, which might be difficult for patients who live in a remote area with poor medical access. To address this issue, we propose an automatic strabismus screening method that calculates the CLR ratio to determine the presence of strabismus based on image processing. In particular, the method first employs a pre-trained face detection model and a 68 facial landmarks detector to extract the eye region image. The data points located in the limbus are then collected, and the least square method is applied to obtain the center coordinates of the iris. Finally, the coordinate of the reflective light point center within the iris is extracted and used to calculate the CLR ratio with the coordinate of iris edges. Experimental results with several images demonstrate that the proposed method can be a promising solution to provide strabismus screening for patients who cannot visit hospitals.

Adaptive Skin Color Segmentation in a Single Image using Image Feedback (영상 피드백을 이용한 단일 영상에서의 적응적 피부색 검출)

  • Do, Jun-Hyeong;Kim, Keun-Ho;Kim, Jong-Yeol
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.3
    • /
    • pp.112-118
    • /
    • 2009
  • Skin color segmentation techniques have been widely utilized for face/hand detection and tracking in many applications such as a diagnosis system using facial information, human-robot interaction, an image retrieval system. In case of a video image, it is common that the skin color model for a target is updated every frame for the robust target tracking against illumination change. As for a single image, however, most of studies employ a fixed skin color model which may result in low detection rate or high false positive errors. In this paper, we propose a novel method for effective skin color segmentation in a single image, which modifies the conditions for skin color segmentation iteratively by the image feedback of segmented skin color region in a given image.

Design of Image Recognition Module for Face and Iris Area based on Pixel with Eye Blinking (눈 깜박임 화소 값 기반의 안면과 홍채영역 영상인식용 모듈설계)

  • Kang, Mingoo
    • Journal of Internet Computing and Services
    • /
    • v.18 no.1
    • /
    • pp.21-26
    • /
    • 2017
  • In this paper, an USB-OTG (Uiversal Serial Bus On-the-go) interface module was designed with the iris information for personal identification. The image recognition algorithm which was searching face and iris areas, was proposed with pixel differences from eye blinking after several facial images were captured and then detected without any activities like as pressing the button of smart phone. The region of pupil and iris could be fast involved with the proper iris area segmentation from the pixel value calculation of frame difference among the images which were detected with two adjacent open-eye and close-eye pictures. This proposed iris recognition could be fast processed with the proper grid size of the eye region, and designed with the frame difference between the adjacent images from the USB-OTG interface with this camera module with the restrict of searching area in face and iris location. As a result, the detection time of iris location can be reduced, and this module can be expected with eliminating the standby time of eye-open.