• Title/Summary/Keyword: 얼굴검출 및 인식

Search Result 211, Processing Time 0.032 seconds

Deep Learning Based Fake Face Detection (딥 러닝 기반의 가짜 얼굴 검출)

  • Kim, DaeHee;Choi, SeungWan;Kwak, SooYeong
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.23 no.5
    • /
    • pp.9-17
    • /
    • 2018
  • Recently, the increasing interest of biometric systems has led to the creation of many researches of biometrics forgery. In order to solve this forgery problem, this paper proposes a method of determining whether a synthesized face made of artificaial intelligence is real face or fake face. The proposed algorithm consists of two steps. Firstly, we create the fake face images using various GAN (Generative Adversarial Networks) algorithms. After that, deep learning algorithm can classify the real face image and the generated face image. The experimental results shows that the proposed algorithm can detect the fake face image which looks like the real face. Also, we obtained the classification accuracy of 88.7%.

A study on local facial features using LDP (LDP를 이용한 지역적 얼굴 특징 표현 방법에 관한 연구)

  • Cho, Young Tak;Jung, Woong Kyung;Ahn, Yong Hak;Chae, Ok Sam
    • Convergence Security Journal
    • /
    • v.14 no.5
    • /
    • pp.49-56
    • /
    • 2014
  • In this paper, we proposed a method for representing local facial features based on LDP (Local Directional Pattern). To represent both PFF (Permanent Facial Features) and TFF (Transient Facial Features) effectively, the proposed method configure local facial feature vectors based on overlapped blocks for each facial feature in the forms of various size and shape. There are three advantages - it take advantages of geometric feature based method; it shows robustness about detection error using movement characteristics of each facial feature; and it shows reduced sampling error because maintain spatial information caused by block size variability. Proposed method shows better classification accuracy and reduced amount of calculation than existing methods.

Robust Real-time Face Detection Scheme on Various illumination Conditions (다양한 조명 환경에 강인한 실시간 얼굴확인 기법)

  • Kim, Soo-Hyun;Han, Young-Joon;Cha, Hyung-Tai;Hahn, Hern-Soo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.7
    • /
    • pp.821-829
    • /
    • 2004
  • A face recognition has been used for verifying and authorizing valid users, but its applications have been restricted according to lighting conditions. In order to minimizing the restricted conditions, this paper proposes a new algorithm of detecting the face from the input image obtained under the irregular lighting condition. First, the proposed algorithm extracts an edge difference image from the input image where a skin color and a face contour are disappeared due to the background color or the lighting direction. In the next step, it extracts a face region using the histogram of the edge difference image and the intensity information. Using the intensity information, the face region is divided into the horizontal regions with feasible facial features. The each of horizontal regions is classified as three groups with the facial features(including eye, nose, and mouth) and the facial features are extracted using empirical properties of the facial features. Only when the facial features satisfy their topological rules, the face region is considered as a face. It has been proved by the experiments that the proposed algorithm can detect faces even when the large portion of face contour is lost due to the inadequate lighting condition or the image background color is similar to the skin color.

Recognition of dog's front face using deep learning and machine learning (딥러닝 및 기계학습 활용 반려견 얼굴 정면판별 방법)

  • Kim, Jong-Bok;Jang, Dong-Hwa;Yang, Kayoung;Kwon, Kyeong-Seok;Kim, Jung-Kon;Lee, Joon-Whoan
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.12
    • /
    • pp.1-9
    • /
    • 2020
  • As pet dogs rapidly increase in number, abandoned and lost dogs are also increasing in number. In Korea, animal registration has been in force since 2014, but the registration rate is not high owing to safety and effectiveness issues. Biometrics is attracting attention as an alternative. In order to increase the recognition rate from biometrics, it is necessary to collect biometric images in the same form as much as possible-from the face. This paper proposes a method to determine whether a dog is facing front or not in a real-time video. The proposed method detects the dog's eyes and nose using deep learning, and extracts five types of directional face information through the relative size and position of the detected face. Then, a machine learning classifier determines whether the dog is facing front or not. We used 2,000 dog images for learning, verification, and testing. YOLOv3 and YOLOv4 were used to detect the eyes and nose, and Multi-layer Perceptron (MLP), Random Forest (RF), and the Support Vector Machine (SVM) were used as classifiers. When YOLOv4 and the RF classifier were used with all five types of the proposed face orientation information, the face recognition rate was best, at 95.25%, and we found that real-time processing is possible.

Intelligent Face Recognition and Tracking System to Distribute GPU Resources using CUDA (쿠다를 사용하여 GPU 리소스를 분배하는 지능형 얼굴 인식 및 트래킹 시스템)

  • Kim, Jae-Heong;Lee, Seung-Ho
    • Journal of IKEEE
    • /
    • v.22 no.2
    • /
    • pp.281-288
    • /
    • 2018
  • In this paper, we propose an intelligent face recognition and tracking system that distributes GPU resources using CUDA. The proposed system consists of five steps such as GPU allocation algorithm that distributes GPU resources in optimal state, face area detection and face recognition using deep learning, real time face tracking, and PTZ camera control. The GPU allocation algorithm that distributes multi-GPU resources optimally distributes the GPU resources flexibly according to the activation level of the GPU, unlike the method of allocating the GPU to the thread fixedly. Thus, there is a feature that enables stable and efficient use of multiple GPUs. In order to evaluate the performance of the proposed system, we compared the proposed system with the non - distributed system. As a result, the system which did not allocate the resource showed unstable operation, but the proposed system showed stable resource utilization because it was operated stably. Thus, the utility of the proposed system has been demonstrated.

3D Facial Animation with Head Motion Estimation and Facial Expression Cloning (얼굴 모션 추정과 표정 복제에 의한 3차원 얼굴 애니메이션)

  • Kwon, Oh-Ryun;Chun, Jun-Chul
    • The KIPS Transactions:PartB
    • /
    • v.14B no.4
    • /
    • pp.311-320
    • /
    • 2007
  • This paper presents vision-based 3D facial expression animation technique and system which provide the robust 3D head pose estimation and real-time facial expression control. Many researches of 3D face animation have been done for the facial expression control itself rather than focusing on 3D head motion tracking. However, the head motion tracking is one of critical issues to be solved for developing realistic facial animation. In this research, we developed an integrated animation system that includes 3D head motion tracking and facial expression control at the same time. The proposed system consists of three major phases: face detection, 3D head motion tracking, and facial expression control. For face detection, with the non-parametric HT skin color model and template matching, we can detect the facial region efficiently from video frame. For 3D head motion tracking, we exploit the cylindrical head model that is projected to the initial head motion template. Given an initial reference template of the face image and the corresponding head motion, the cylindrical head model is created and the foil head motion is traced based on the optical flow method. For the facial expression cloning we utilize the feature-based method, The major facial feature points are detected by the geometry of information of the face with template matching and traced by optical flow. Since the locations of varying feature points are composed of head motion and facial expression information, the animation parameters which describe the variation of the facial features are acquired from geometrically transformed frontal head pose image. Finally, the facial expression cloning is done by two fitting process. The control points of the 3D model are varied applying the animation parameters to the face model, and the non-feature points around the control points are changed by use of Radial Basis Function(RBF). From the experiment, we can prove that the developed vision-based animation system can create realistic facial animation with robust head pose estimation and facial variation from input video image.

Design of Real-time Face Recognition Systems Based on Data-Preprocessing and Neuro-Fuzzy Networks for the Improvement of Recognition Rate (인식률 향상을 위한 데이터 전처리와 Neuro-Fuzzy 네트워크 기반의 실시간 얼굴 인식 시스템 설계)

  • Yoo, Sung-Hoon;Oh, Sung-Kwun;Kim, Hyun-Ki
    • Proceedings of the KIEE Conference
    • /
    • 2011.07a
    • /
    • pp.1952-1953
    • /
    • 2011
  • 본 논문에서는 다항식 기반 Radial Basis Function(RBF)신경회로망(Polynomial based Radial Basis function Neural Network)을 설계하고 이를 n-클래스 패턴 분류 문제에 적용한다. 제안된 다항식기반 RBF 신경회로망은 입력층, 은닉층, 출력층으로 이루어진다. 입력층은 입력 벡터의 값들을 은닉층으로 전달하는 기능을 수행하고 은닉층과 출력층사이의 연결가중치는 상수, 선형식 또는 이차식으로 이루어지며 경사 하강법에 의해 학습된다. Networks의 최종 출력은 연결가중치와 은닉층 출력의 곱에 의해 퍼지추론의 결과로서 얻어진다. 패턴분류기의 최적화는 PSO(Particle Swarm Optimization)알고리즘을 통해 이루어진다. 그리고 제안된 패턴분류기는 실제 얼굴인식 시스템으로 응용하여 직접 CCD 카메라로부터 입력받은 데이터를 영상 보정, 얼굴 검출, 특징 추출 등과 같은 처리 과정을 포함하여 서로 다른 등록인물의 n-클래스 분류 문제에 적용 및 평가되어 분류기로써의 성능을 분석해본다.

  • PDF

Face recognition rate comparison with distance change using embedded data in stereo images (스테레오 영상에서 임베디드 데이터를 이용한 거리에 따른 얼굴인식률 비교)

  • 박장한;남궁재찬
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.6
    • /
    • pp.81-89
    • /
    • 2004
  • In this paper, we compare face recognition rate by PCA algorithm using distance change and embedded data being input left side and right side image in stereo images. The proposed method detects face region from RGB color space to YCbCr color space. Also, The extracted face image's scale up/down according to distance change and extracts more robust face region. The proposed method through an experiment could establish standard distance (100cm) in distance about 30∼200cm, and get 99.05% (100cm) as an average recognition result by scale change. The definition of super state is specification region in normalized size (92${\times}$112), and the embedded data extracts the inner factor of defined super state, achieved face recognition through PCA algorithm. The orignal images can receive specification data in limited image's size (92${\times}$112) because embedded data to do learning not that do all learning, in image of 92${\times}$112 size averagely 99.05%, shows face recognition rate of test 1 99.05%, test 2 98.93%, test 3 98.54%, test 4 97.85%. Therefore, the proposed method through an experiment showed that if apply distance change rate could get high recognition rate, and the processing speed improved as well as reduce face information.

Estimation of a Driver's Physical Condition Using Real-time Vision System (실시간 비전 시스템을 이용한 운전자 신체적 상태 추정)

  • Kim, Jong-Il;Ahn, Hyun-Sik;Jeong, Gu-Min;Moon, Chan-Woo
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.9 no.5
    • /
    • pp.213-224
    • /
    • 2009
  • This paper presents a new algorithm for estimating a driver's physical condition using real-time vision system and performs experimentation for real facial image data. The system relies on a face recognition to robustly track the center points and sizes of person's two pupils, and two side edge points of the mouth. The face recognition constitutes the color statistics by YUV color space together with geometrical model of a typical face. The system can classify the rotation in all viewing directions, to detect eye/mouth occlusion, eye blinking and eye closure, and to recover the three dimensional gaze of the eyes. These are utilized to determine the carelessness and drowsiness of the driver. Finally, experimental results have demonstrated the validity and the applicability of the proposed method for the estimation of a driver's physical condition.

  • PDF

Driver drowsiness recognition system based on camera image analysis (카메라 영상 분석 기반 운전자 졸음 인식 시스템)

  • Kim, Hyun-Suk;Choi, Min-Su;Bae, You-Suk
    • Annual Conference of KIPS
    • /
    • 2016.04a
    • /
    • pp.719-722
    • /
    • 2016
  • 운전자의 주의력 감쇠는 교통사고 요인에 있어서 큰 비중을 차지한다. 주의력 감쇠는 무선 통화, 기기 조작, 졸음으로 나타날 수 있는데 자동차 대형사고의 대부분은 졸음운전으로 인하여 일어나며, 졸음운전 시에는 운전자의 운전조작 및 방어 조작 능력이 현저하게 저하한다. 본 시스템은 카메라로부터 실시간으로 영상 데이터를 입력 받아 처리하여 운전자의 졸음 상태를 인식하는 시스템으로 운전자에게 졸음방지 기능을 제공한다. Haar-Like Feature cascade classifier 방법을 사용하여 얼굴 및 눈 영역 검출을 하였고 Open Eye, Closed Eye가 학습된 MLP(Multi-Layer Perceptron)를 이용해 눈 깜박임을 인식하여 PERCLOS(Percentage of Eye Close)방법으로 졸음을 판단하였다. 본 논문에서 제안한 방법의 인식률의 정확도를 검증하기 위해 인식률 테스트를 하였다.