• Title/Summary/Keyword: Head tracking

Search Result 245, Processing Time 0.024 seconds

Gaze Detection by Wearable Eye-Tracking and NIR LED-Based Head-Tracking Device Based on SVR

  • Cho, Chul Woo;Lee, Ji Woo;Shin, Kwang Yong;Lee, Eui Chul;Park, Kang Ryoung;Lee, Heekyung;Cha, Jihun
    • ETRI Journal
    • /
    • v.34 no.4
    • /
    • pp.542-552
    • /
    • 2012
  • In this paper, a gaze estimation method is proposed for use with a large-sized display at a distance. Our research has the following four novelties: this is the first study on gaze-tracking for large-sized displays and large Z (viewing) distances; our gaze-tracking accuracy is not affected by head movements since the proposed method tracks the head by using a near infrared camera and an infrared light-emitting diode; the threshold for local binarization of the pupil area is adaptively determined by using a p-tile method based on circular edge detection irrespective of the eyelid or eyelash shadows; and accurate gaze position is calculated by using two support vector regressions without complicated calibrations for the camera, display, and user's eyes, in which the gaze positions and head movements are used as feature values. The root mean square error of gaze detection is calculated as $0.79^{\circ}$ for a 30-inch screen.

Tracking by Detection of Multiple Faces using SSD and CNN Features

  • Tai, Do Nhu;Kim, Soo-Hyung;Lee, Guee-Sang;Yang, Hyung-Jeong;Na, In-Seop;Oh, A-Ran
    • Smart Media Journal
    • /
    • v.7 no.4
    • /
    • pp.61-69
    • /
    • 2018
  • Multi-tracking of general objects and specific faces is an important topic in the field of computer vision applicable to many branches of industry such as biometrics, security, etc. The rapid development of deep neural networks has resulted in a dramatic improvement in face recognition and object detection problems, which helps improve the multiple-face tracking techniques exploiting the tracking-by-detection method. Our proposed method uses face detection trained with a head dataset to resolve the face deformation problem in the tracking process. Further, we use robust face features extracted from the deep face recognition network to match the tracklets with tracking faces using Hungarian matching method. We achieved promising results regarding the usage of deep face features and head detection in a face tracking benchmark.

Neural Network Compensation for Improvement of Real-Time Moving Object Tracking Performance of the ROBOKER Head with a Virtual Link (가상링크 기반의 ROBOKER 머리의 실시간 대상체 추종 성능 향상을 위한 신경망 제어)

  • Kim, Dong-Min;Choi, Ho-Jin;Lee, Geun-Hyung;Jung, Seul
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.15 no.7
    • /
    • pp.694-699
    • /
    • 2009
  • This paper presents the implementation of the real-time object tracking control of the ROBOKER head. The visual servoing technique is used to track the moving object, but suffers from ill-estimated Jacobian of the virtual link design. To improve the tracking performance, the RBF(Radial Basis Function) network is used to compensate for uncertainties in the kinematics of the robot head in on-line fashion. The reference compensation technique is employed as a neural network control scheme. Performances of three schemes, the kinematic based scheme, the Jacobian based scheme, and the neural network compensation scheme are verified by experimental studies. The neural compensation scheme performs best.

Vision-based recognition of a simple non-verbal intent representation by head movements (고개운동에 의한 단순 비언어 의사표현의 비전인식)

  • Yu, Gi-Ho;No, Deok-Su;Lee, Seong-Cheol
    • Journal of the Ergonomics Society of Korea
    • /
    • v.19 no.1
    • /
    • pp.91-100
    • /
    • 2000
  • In this paper the intent recognition system which recognizes the human's head movements as a simple non-verbal intent representation is presented. The system recognizes five basic intent representations. i.e., strong/weak affirmation. strong/weak negation, and ambiguity by image processing of nodding or shaking movements of head. The vision system for tracking the head movements is composed of CCD camera, image processing board and personal computer. The modified template matching method which replaces the reference image with the searched target image in the previous step is used for the robust tracking of the head movements. For the improvement of the processing speed, the searching is performed in the pyramid representation of the original image. By inspecting the variance of the head movement trajectories. we can recognizes the two basic intent representations - affirmation and negation. Also, by focusing the speed of the head movements, we can see the possibility which recognizes the strength of the intent representation.

  • PDF

Effect on Audio Play Latency for Real-Time HMD-Based Headphone Listening (HMD를 이용한 오디오 재생 기술에서 Latency의 영향 분석)

  • Son, Sangmo;Jo, Hyun;Kim, Sunmin
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2014.10a
    • /
    • pp.141-145
    • /
    • 2014
  • A minimally appropriate time delay of audio data processing is investigated for rendering virtual sound source direction in real-time head-tracking environment under headphone listening. Less than 3.7 degree of angular mismatch should be maintained in order to keep desired sound source directions in virtually fixed while listeners are rotating their head in a horizontal plane. The angular mismatch is proportional to speed of head rotation and data processing delay. For 20 degree/s head rotation, which is a relatively slow head-movement case, less than total of 63ms data processing delay should be considered.

  • PDF

Developing Head/Eye Tracking System and Sync Verification (헤드/아이 통합 트랙커 개발 및 통합 성능 검증)

  • Kim, Jeong-Ho;Lee, Dae-Woo;Heo, Se-Jong;Park, Chan-Gook;Baek, Kwang-Yul;Bang, Hyo-Choong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.1
    • /
    • pp.90-95
    • /
    • 2010
  • This paper describes the development of integrated head and eye tracker system. Vision based head tracker is performed and it has 7mm error in 300mm translation. The epi-polar method and point matching are used for determining a position of head and rotational degree. High brightness LEDs are installed on helmet and the installed pattern is very important to match the points of stereo system. Eye tracker also uses LED for constant illumination. A Position of gazed object(3m distance) is determined by pupil tracking and eye tracker has 1~5 pixel error. Integration of result data of each tracking system is important. RS-232C communication is applied to integrated system and triggering signal is used for synchronization.

3D Facial Landmark Tracking and Facial Expression Recognition

  • Medioni, Gerard;Choi, Jongmoo;Labeau, Matthieu;Leksut, Jatuporn Toy;Meng, Lingchao
    • Journal of information and communication convergence engineering
    • /
    • v.11 no.3
    • /
    • pp.207-215
    • /
    • 2013
  • In this paper, we address the challenging computer vision problem of obtaining a reliable facial expression analysis from a naturally interacting person. We propose a system that combines a 3D generic face model, 3D head tracking, and 2D tracker to track facial landmarks and recognize expressions. First, we extract facial landmarks from a neutral frontal face, and then we deform a 3D generic face to fit the input face. Next, we use our real-time 3D head tracking module to track a person's head in 3D and predict facial landmark positions in 2D using the projection from the updated 3D face model. Finally, we use tracked 2D landmarks to update the 3D landmarks. This integrated tracking loop enables efficient tracking of the non-rigid parts of a face in the presence of large 3D head motion. We conducted experiments for facial expression recognition using both framebased and sequence-based approaches. Our method provides a 75.9% recognition rate in 8 subjects with 7 key expressions. Our approach provides a considerable step forward toward new applications including human-computer interactions, behavioral science, robotics, and game applications.

Active Facial Tracking for Fatigue Detection (피로 검출을 위한 능동적 얼굴 추적)

  • Kim, Tae-Woo;Kang, Yong-Seok
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.2 no.3
    • /
    • pp.53-60
    • /
    • 2009
  • The vision-based driver fatigue detection is one of the most prospective commercial applications of facial expression recognition technology. The facial feature tracking is the primary technique issue in it. Current facial tracking technology faces three challenges: (1) detection failure of some or all of features due to a variety of lighting conditions and head motions; (2) multiple and non-rigid object tracking; and (3) features occlusion when the head is in oblique angles. In this paper, we propose a new active approach. First, the active IR sensor is used to robustly detect pupils under variable lighting conditions. The detected pupils are then used to predict the head motion. Furthermore, face movement is assumed to be locally smooth so that a facial feature can be tracked with a Kalman filter. The simultaneous use of the pupil constraint and the Kalman filtering greatly increases the prediction accuracy for each feature position. Feature detection is accomplished in the Gabor space with respect to the vicinity of predicted location. Local graphs consisting of identified features are extracted and used to capture the spatial relationship among detected features. Finally, a graph-based reliability propagation is proposed to tackle the occlusion problem and verify the tracking results. The experimental results show validity of our active approach to real-life facial tracking under variable lighting conditions, head orientations, and facial expressions.

  • PDF

Active Facial Tracking for Fatigue Detection (피로 검출을 위한 능동적 얼굴 추적)

  • 박호식;정연숙;손동주;나상동;배철수
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2004.05b
    • /
    • pp.603-607
    • /
    • 2004
  • The vision-based driver fatigue detection is one of the most prospective commercial applications of facial expression recognition technology. The facial feature tracking is the primary technique issue in it. Current facial tracking technology faces three challenges: (1) detection failure of some or all of features due to a variety of lighting conditions and head motions; (2) multiple and non-rigid object tracking and (3) features occlusion when the head is in oblique angles. In this paper, we propose a new active approach. First, the active IR sensor is used to robustly detect pupils under variable lighting conditions. The detected pupils are then used to predict the head motion. Furthermore, face movement is assumed to be locally smooth so that a facial feature can be tracked with a Kalman filter. The simultaneous use of the pupil constraint and the Kalman filtering greatly increases the prediction accuracy for each feature position. Feature detection is accomplished in the Gabor space with respect to the vicinity of predicted location. Local graphs consisting of identified features are extracted and used to capture the spatial relationship among detected features. Finally, a graph-based reliability propagation is proposed to tackle the occlusion problem and verify the tracking results. The experimental results show validity of our active approach to real-life facial tracking under variable lighting conditions, head orientations, and facial expressions.

  • PDF

The depth quality enhancement algorithm for Autostereoscopic 3D Monitor (무안경 3D 모니터를 위한 Depth 화질 향상 Algorithm)

  • Song, Sung-Ho;Lee, Kyoung-Il;Lee, Dong-Ha;Park, Jong-Cheol;Lee, Jea-Jun;Kim, Young-Kil
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2012.05a
    • /
    • pp.133-136
    • /
    • 2012
  • In this paper, we found the many effective ways and apply for improve the 3D quality of Autostereoscopic 3D display products. Autostereoscopic products compared to traditional 3D glasses, the disadvantage is the poor depth of 3D picture quality and it only can see the fixed distance and position. So, for the compensate this disadvantage, we use the Head tracking technology and video placement algorithms and several techniques. In this paper, the will report on how to improve the Parallax Barrier Autostereoscopic 3D quality through the Head tracking of the user identification, video replacement algorithms and crosstalk improving method.

  • PDF