• Title/Summary/Keyword: camera pose

Search Result 271, Processing Time 0.029 seconds

MPEG-U based Advanced User Interaction Interface System Using Hand Posture Recognition (손 자세 인식을 이용한 MPEG-U 기반 향상된 사용자 상호작용 인터페이스 시스템)

  • Han, Gukhee;Lee, Injae;Choi, Haechul
    • Journal of Broadcast Engineering
    • /
    • v.19 no.1
    • /
    • pp.83-95
    • /
    • 2014
  • Hand posture recognition is an important technique to enable a natural and familiar interface in HCI(human computer interaction) field. In this paper, we introduce a hand posture recognition method by using a depth camera. Moreover, the hand posture recognition method is incorporated with MPEG-U based advanced user interaction (AUI) interface system, which can provide a natural interface with a variety of devices. The proposed method initially detects positions and lengths of all fingers opened and then it recognizes hand posture from pose of one or two hands and the number of fingers folded when user takes a gesture representing a pattern of AUI data format specified in the MPEG-U part 2. The AUI interface system represents user's hand posture as compliant MPEG-U schema structure. Experimental results show performance of the hand posture recognition and it is verified that the AUI interface system is compatible with the MPEG-U standard.

Design of a Background Image Based Multi-Degree-of-Freedom Pointing Device (배경영상 기반 다자유도 포인팅 디바이스의 설계)

  • Jang, Suk-Yoon;Kho, Jae-Won
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.6
    • /
    • pp.133-141
    • /
    • 2008
  • As interactive multimedia have come into wide use, user interfaces such as remote controllers or classical computer mice have several limitations that cause inconvenience. We propose a vision-based pointing device to resolve this problem. We analyzed the moving image from the camera which is embedded in the pointing device and estimate the movement of the device. The pose of the cursor can be determined from this result. To process in the real time, we used the low resolution of $288{\times}208$ pixel camera and comer points of the screen were tracked using local optical flow method. The distance from screen and device was calculated from the size of screen in the image. The proposed device has simple configurations, low cost, easy use, and intuitive handhold operation like traditional mice. Moreover it shows reliable performance even in the dark condition.

1-Point Ransac Based Robust Visual Odometry

  • Nguyen, Van Cuong;Heo, Moon Beom;Jee, Gyu-In
    • Journal of Positioning, Navigation, and Timing
    • /
    • v.2 no.1
    • /
    • pp.81-89
    • /
    • 2013
  • Many of the current visual odometry algorithms suffer from some extreme limitations such as requiring a high amount of computation time, complex algorithms, and not working in urban environments. In this paper, we present an approach that can solve all the above problems using a single camera. Using a planar motion assumption and Ackermann's principle of motion, we construct the vehicle's motion model as a circular planar motion (2DOF). Then, we adopt a 1-point method to improve the Ransac algorithm and the relative motion estimation. In the Ransac algorithm, we use a 1-point method to generate the hypothesis and then adopt the Levenberg-Marquardt method to minimize the geometric error function and verify inliers. In motion estimation, we combine the 1-point method with a simple least-square minimization solution to handle cases in which only a few feature points are present. The 1-point method is the key to speed up our visual odometry application to real-time systems. Finally, a Bundle Adjustment algorithm is adopted to refine the pose estimation. The results on real datasets in urban dynamic environments demonstrate the effectiveness of our proposed algorithm.

Performance Analysis of Feature Detection Methods for Topology-Based Feature Description (토폴로지 기반 특징 기술을 위한 특징 검출 방법의 성능 분석)

  • Park, Han-Hoon;Moon, Kwang-Seok
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.16 no.2
    • /
    • pp.44-49
    • /
    • 2015
  • When the scene has less texture or when camera pose largely changes, the existing texture-based feature tracking methods are not reliable. Topology-based feature description methods, which use the geometric relationship between features such as LLAH, is a good alternative. However, they require feature detection methods with high performance. As a basic study on developing an effective feature detection method for topology-based feature description, this paper aims at examining their applicability to topology-based feature description by analyzing the repeatability of several feature detection methods that are included in the OpenCV library. Experimental results show that FAST outperforms the others.

Absolute Positioning System for Mobile Robot Navigation in an Indoor Environment (ICCAS 2004)

  • Yun, Jae-Mu;Park, Jin-Woo;Choi, Ho-Seek;Lee, Jang-Myung
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.1448-1451
    • /
    • 2004
  • Position estimation is one of the most important functions for the mobile robot navigating in the unstructured environment. Most of previous localization schemes estimate current position and pose of mobile robot by applying various localization algorithms with the information obtained from sensors which are set on the mobile robot, or by recognizing an artificial landmark attached on the wall, or objects of the environment as natural landmark in the indoor environment. Several drawbacks about them have been brought up. To compensate the drawbacks, a new localization method that estimates the absolute position of the mobile robot by using a fixed camera on the ceiling in the corridor is proposed. And also, it can improve the success rate for position estimation using the proposed method, which calculates the real size of an object. This scheme is not a relative localization, which decreases the position error through algorithms with noisy sensor data, but a kind of absolute localization. The effectiveness of the proposed localization scheme is demonstrated through the experiments.

  • PDF

Point Pattern Matching Based Global Localization using Ceiling Vision (천장 조명을 이용한 점 패턴 매칭 기반의 광역적인 위치 추정)

  • Kang, Min-Tae;Sung, Chang-Hun;Roh, Hyun-Chul;Chung, Myung-Jin
    • Proceedings of the KIEE Conference
    • /
    • 2011.07a
    • /
    • pp.1934-1935
    • /
    • 2011
  • In order for a service robot to perform several tasks, basically autonomous navigation technique such as localization, mapping, and path planning is required. The localization (estimation robot's pose) is fundamental ability for service robot to navigate autonomously. In this paper, we propose a new system for point pattern matching based visual global localization using spot lightings in ceiling. The proposed algorithm us suitable for system that demands high accuracy and fast update rate such a guide robot in the exhibition. A single camera looking upward direction (called ceiling vision system) is mounted on the head of the mobile robot and image features such as lightings are detected and tracked through the image sequence. For detecting more spot lightings, we choose wide FOV lens, and inevitably there is serious image distortion. But by applying correction calculation only for the position of spot lightings not whole image pixels, we can decrease the processing time. And then using point pattern matching and least square estimation, finally we can get the precise position and orientation of the mobile robot. Experimental results demonstrate the accuracy and update rate of the proposed algorithm in real environments.

  • PDF

Performance Optimization of LLAH for Tracking Random Dots under Gaussian Noise (가우시안 잡음을 가지는 랜덤 점 추적을 위한 LLAH의 성능 최적화)

  • Park, Hanhoon
    • Journal of Broadcast Engineering
    • /
    • v.20 no.6
    • /
    • pp.912-920
    • /
    • 2015
  • Unlike general texture-based feature description algorithms, Locally Likely Arrangement Hashing (LLAH) algorithm describes a feature based on the geometric relationship between its neighbors. Thus, even in poor-textured scenes or large camera pose changes, it can successfully describe and track features and enables to implement augmented reality. This paper aims to optimize the performance of LLAH algorithm for tracking random dots (= features) with Gaussian noise. For this purpose, images with different number of features and magnitude of Gaussian noise are prepared. Then, the performance of LLAH algorithm according to the conditions: the number of neighbors, the type of geometric invariants, and the distance between features, is analyzed, and the optimal conditions are determined. With the optimal conditions, each feature could be matched and tracked in real-time with a matching rate of more than 80%.

Detecting Complex 3D Human Motions with Body Model Low-Rank Representation for Real-Time Smart Activity Monitoring System

  • Jalal, Ahmad;Kamal, Shaharyar;Kim, Dong-Seong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.3
    • /
    • pp.1189-1204
    • /
    • 2018
  • Detecting and capturing 3D human structures from the intensity-based image sequences is an inherently arguable problem, which attracted attention of several researchers especially in real-time activity recognition (Real-AR). These Real-AR systems have been significantly enhanced by using depth intensity sensors that gives maximum information, in spite of the fact that conventional Real-AR systems are using RGB video sensors. This study proposed a depth-based routine-logging Real-AR system to identify the daily human activity routines and to make these surroundings an intelligent living space. Our real-time routine-logging Real-AR system is categorized into two categories. The data collection with the use of a depth camera, feature extraction based on joint information and training/recognition of each activity. In-addition, the recognition mechanism locates, and pinpoints the learned activities and induces routine-logs. The evaluation applied on the depth datasets (self-annotated and MSRAction3D datasets) demonstrated that proposed system can achieve better recognition rates and robust as compare to state-of-the-art methods. Our Real-AR should be feasibly accessible and permanently used in behavior monitoring applications, humanoid-robot systems and e-medical therapy systems.

Hand Gesture Interface for Manipulating 3D Objects in Augmented Reality (증강현실에서 3D 객체 조작을 위한 손동작 인터페이스)

  • Park, Keon-Hee;Lee, Guee-Sang
    • The Journal of the Korea Contents Association
    • /
    • v.10 no.5
    • /
    • pp.20-28
    • /
    • 2010
  • In this paper, we propose a hand gesture interface for the manipulation of augmented objects in 3D space using a camera. Generally a marker is used for the detection of 3D movement in 2D images. However marker based system has obvious defects since markers are always to be included in the image or we need additional equipments for controling objects, which results in reduced immersion. To overcome this problem, we replace marker by planar hand shape by estimating the hand pose. Kalman filter is for robust tracking of the hand shape. The experimental result indicates the feasibility of the proposed algorithm for hand based AR interfaces.

ERGONOMIC ANALYSIS OF A TELEMANIPULATION TECHNIQUE FOR A PYROPROCESS DEMONSTRATION FACILITY

  • Yu, Seungnam;Lee, Jongkwang;Park, Byungsuk;Kim, Kiho;Cho, Ilje
    • Nuclear Engineering and Technology
    • /
    • v.46 no.4
    • /
    • pp.489-500
    • /
    • 2014
  • In this study, remote handling strategies for a large-scale argon cell facility were considered. The suggested strategies were evaluated by several types of field test. The teleoperation tasks were performed using a developed remote handling system, which enabled traveling over entire cell area using a bridge transport system. Each arm of the system had six DOFs (degrees of freedom), and the bridge transport system had four DOFs. However, despite the dexterous manipulators and redundant monitoring system, many operators, including professionals, experienced difficulties in operating the remote handling system. This was because of the lack of a strategy for handling the installed camera system, and the difficulty in recognizing the gripper pose, which might fall outside the FOV (field of vision) of the system during teleoperation. Hence, in this paper, several considerations for the remote handling tasks performed in the target facility were discussed, and the tasks were analyzed based on ergonomic factors such as the workload. Toward the development of a successful operation strategy, several ergonomic issues, such as active/passive view of the remote handling system, eye/hand alignment, and FOV were considered. Furthermore, using the method for classifying remote handling tasks, several unit tasks were defined and evaluated.