• Title/Summary/Keyword: Pose estimation

Search Result 395, Processing Time 0.022 seconds

Head Pose Estimation Based on Perspective Projection Using PTZ Camera (원근투영법 기반의 PTZ 카메라를 이용한 머리자세 추정)

  • Kim, Jin Suh;Lee, Gyung Ju;Kim, Gye Young
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.7 no.7
    • /
    • pp.267-274
    • /
    • 2018
  • This paper describes a head pose estimation method using PTZ(Pan-Tilt-Zoom) camera. When the external parameters of a camera is changed by rotation and translation, the estimated face pose for the same head also varies. In this paper, we propose a new method to estimate the head pose independently on varying the parameters of PTZ camera. The proposed method consists of 3 steps: face detection, feature extraction, and pose estimation. For each step, we respectively use MCT(Modified Census Transform) feature, the facial regression tree method, and the POSIT(Pose from Orthography and Scaling with ITeration) algorithm. The existing POSIT algorithm does not consider the rotation of a camera, but this paper improves the POSIT based on perspective projection in order to estimate the head pose robustly even when the external parameters of a camera are changed. Through experiments, we confirmed that RMSE(Root Mean Square Error) of the proposed method improve $0.6^{\circ}$ less then the conventional method.

2D - 3D Human Face Verification System based on Multiple RGB-D Camera using Head Pose Estimation (얼굴 포즈 추정을 이용한 다중 RGB-D 카메라 기반의 2D - 3D 얼굴 인증을 위한 시스템)

  • Kim, Jung-Min;Li, Shengzhe;Kim, Hak-Il
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.24 no.4
    • /
    • pp.607-616
    • /
    • 2014
  • Face recognition is a big challenge in surveillance system since different rotation angles of the face make the difficulty to recognize the face of the same person. This paper proposes a novel method to recognize face with different head poses by using 3D information of the face. Firstly, head pose estimation (estimation of different head pose angles) is accomplished by the POSIT algorithm. Then, 3D face image data is constructed by using head pose estimation. After that, 2D image and the constructed 3D face matching is performed. Face verification is accomplished by using commercial face recognition SDK. Performance evaluation of the proposed method indicates that the error range of head pose estimation is below 10 degree and the matching rate is about 95%.

Development of 3-Dimensional Pose Estimation Algorithm using Inertial Sensors for Humanoid Robot (관성 센서를 이용한 휴머노이드 로봇용 3축 자세 추정 알고리듬 개발)

  • Lee, Ah-Lam;Kim, Jung-Han
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.14 no.2
    • /
    • pp.133-140
    • /
    • 2008
  • In this paper, a small and effective attitude estimation system for a humanoid robot was developed. Four small inertial sensors were packed and used for inertial measurements(3D accelerometer and three 1D gyroscopes.) An effective 3D pose estimation algorithm for low cost DSP using an extended Kalman filter was developed and evaluated. The 3D pose estimation algorithm has a very simple structure composed by 3 modules of a linear acceleration estimator, an external acceleration detector and an pseudo-accelerometer output estimator. The algorithm also has an effective switching structure based on probability and simple feedback loop for the extended Kalman filter. A special test equipment using linear motor for the testing of the 3D pose sensor was developed and the experimental results showed its very fast convergence to real values and effective responses. Popular DSP of TMS320F2812 was used to calculate robot's 3D attitude and translated acceleration, and the whole system were packed in a small size for humanoids robots. The output of the 3D sensors(pitch, roll, 3D linear acceleration, and 3D angular rate) can be transmitted to a humanoid robot at 200Hz frequency.

A Kidnapping Detection Using Human Pose Estimation in Intelligent Video Surveillance Systems

  • Park, Ju Hyun;Song, KwangHo;Kim, Yoo-Sung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.8
    • /
    • pp.9-16
    • /
    • 2018
  • In this paper, a kidnapping detection scheme in which human pose estimation is used to classify accurately between kidnapping cases and normal ones is proposed. To estimate human poses from input video, human's 10 joint information is extracted by OpenPose library. In addition to the features which are used in the previous study to represent the size change rates and the regularities of human activities, the human pose estimation features which are computed from the location of detected human's joints are used as the features to distinguish kidnapping situations from the normal accompanying ones. A frame-based kidnapping detection scheme is generated according to the selection of J48 decision tree model from the comparison of several representative classification models. When a video has more frames of kidnapping situation than the threshold ratio after two people meet in the video, the proposed scheme detects and notifies the occurrence of kidnapping event. To check the feasibility of the proposed scheme, the detection accuracy of our newly proposed scheme is compared with that of the previous scheme. According to the experiment results, the proposed scheme could detect kidnapping situations more 4.73% correctly than the previous scheme.

Design of Face Recognition System Based on Pose Estimation : Comparative Studies of Pose Estimation Algorithms (포즈 추정 기반 얼굴 인식 시스템 설계 : 포즈 추정 알고리즘 비교 연구)

  • Kim, Jin-Yul;Kim, Jong-Bum;Oh, Sung-Kwun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.66 no.4
    • /
    • pp.672-681
    • /
    • 2017
  • This paper is concerned with the design methodology of face recognition system based on pose estimation. In 2-dimensional face recognition, the variations of facial pose cause the deterioration of recognition performance because object recognition is carried out by using brightness of each pixel on image. To alleviate such problem, the proposed face recognition system deals with Learning Vector Quantizatioin(LVQ) or K-Nearest Neighbor(K-NN) to estimate facial pose on image and then the images obtained from LVQ or K-NN are used as the inputs of networks such as Convolution Neural Networks(CNNs) and Radial Basis Function Neural Networks(RBFNNs). The effectiveness and efficiency of the post estimation using LVQ and K-NN as well as face recognition rate using CNNs and RBFNNs are discussed through experiments carried out by using ICPR and CMU PIE databases.

Analysis of Pitching Motions by Human Pose Estimation Based on RGB Images (RGB 이미지 기반 인간 동작 추정을 통한 투구 동작 분석)

  • Yeong Ju Woo;Ji-Yong Joo;Young-Kwan Kim;Hie Yong Jeong
    • Smart Media Journal
    • /
    • v.13 no.4
    • /
    • pp.16-22
    • /
    • 2024
  • Pitching is a major part of baseball, so much so that it can be said to be the beginning of baseball. Analysis of accurate pitching motions is very important in terms of performance improvement and injury prevention. When analyzing the correct pitching motion, the currently used motion capture method has several critical environmental drawbacks. In this paper, we propose analysis of pitching motion using the RGB-based Human Pose Estimation (HPE) model to replace motion capture, which has these shortcomings, and use motion capture data and HPE data to verify its reliability. The similarity of the two data was verified by comparing joint coordinates using the Dynamic Time Warping (DTW) algorithm.

Pose Tracking of Moving Sensor using Monocular Camera and IMU Sensor

  • Jung, Sukwoo;Park, Seho;Lee, KyungTaek
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.8
    • /
    • pp.3011-3024
    • /
    • 2021
  • Pose estimation of the sensor is important issue in many applications such as robotics, navigation, tracking, and Augmented Reality. This paper proposes visual-inertial integration system appropriate for dynamically moving condition of the sensor. The orientation estimated from Inertial Measurement Unit (IMU) sensor is used to calculate the essential matrix based on the intrinsic parameters of the camera. Using the epipolar geometry, the outliers of the feature point matching are eliminated in the image sequences. The pose of the sensor can be obtained from the feature point matching. The use of IMU sensor can help initially eliminate erroneous point matches in the image of dynamic scene. After the outliers are removed from the feature points, these selected feature points matching relations are used to calculate the precise fundamental matrix. Finally, with the feature point matching relation, the pose of the sensor is estimated. The proposed procedure was implemented and tested, comparing with the existing methods. Experimental results have shown the effectiveness of the technique proposed in this paper.

A Multi-Stage Convolution Machine with Scaling and Dilation for Human Pose Estimation

  • Nie, Yali;Lee, Jaehwan;Yoon, Sook;Park, Dong Sun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.6
    • /
    • pp.3182-3198
    • /
    • 2019
  • Vision-based Human Pose Estimation has been considered as one of challenging research subjects due to problems including confounding background clutter, diversity of human appearances and illumination changes in scenes. To tackle these problems, we propose to use a new multi-stage convolution machine for estimating human pose. To provide better heatmap prediction of body joints, the proposed machine repeatedly produces multiple predictions according to stages with receptive field large enough for learning the long-range spatial relationship. And stages are composed of various modules according to their strategic purposes. Pyramid stacking module and dilation module are used to handle problem of human pose at multiple scales. Their multi-scale information from different receptive fields are fused with concatenation, which can catch more contextual information from different features. And spatial and channel information of a given input are converted to gating factors by squeezing the feature maps to a single numeric value based on its importance in order to give each of the network channels different weights. Compared with other ConvNet-based architectures, we demonstrated that our proposed architecture achieved higher accuracy on experiments using standard benchmarks of LSP and MPII pose datasets.

A Method for Improving Accuracy of Object Recognition and Pose Estimation by Using Kinect sensor (Kinect센서를 이용한 물체 인식 및 자세 추정을 위한 정확도 개선 방법)

  • Kim, Anna;Yee, Gun Kyu;Kang, Gitae;Kim, Yong Bum;Choi, Hyouk Ryeol
    • The Journal of Korea Robotics Society
    • /
    • v.10 no.1
    • /
    • pp.16-23
    • /
    • 2015
  • This paper presents a method of improving the pose recognition accuracy of objects by using Kinect sensor. First, by using the SURF algorithm, which is one of the most widely used local features point algorithms, we modify inner parameters of the algorithm for efficient object recognition. The proposed method is adjusting the distance between the box filter, modifying Hessian matrix, and eliminating improper key points. In the second, the object orientation is estimated based on the homography. Finally the novel approach of Auto-scaling method is proposed to improve accuracy of object pose estimation. The proposed algorithm is experimentally tested with objects in the plane and its effectiveness is validated.

Pose Estimation of an Object from X-ray Images Based on Principal Axis Analysis

  • Roh, Young-Jun;Cho, Hyung-Suck
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2002.10a
    • /
    • pp.97.4-97
    • /
    • 2002
  • 1. Introduction Pose estimation of a three dimensional object has been studied in robot vision area, and it is needed in a number of industrial applications such as process monitoring and control, assembly and PCB inspection. In this research, we propose a new pose estimation method based on principal axes analysis. Here, it is assumed that the locations of x-ray source and the image plane are predetermined and the object geometry is known. To this end, we define a dispersion matrix of an object, which is a discrete form of inertia matrix of the object. It can be determined here from a set of x-ray images, at least three images are required. Then, the pose information is obtained fro...

  • PDF