• Title/Summary/Keyword: camera pose

Search Result 271, Processing Time 0.035 seconds

The Spreader Pose Determination Research Using CCD Camera and Laser Range Finder (CCD카메라와 레이저 거리미터기를 이용한 스프레더 자세 인식 방법 연구)

  • 이봉기;박수민;진태석;이장명;이권순
    • Proceedings of the Korean Institute of Navigation and Port Research Conference
    • /
    • 2002.11a
    • /
    • pp.121-126
    • /
    • 2002
  • This paper introduces a method that can get information for the movement of a spreader and skew in order to drive ALS(Automatic Landing System) in the crane used at the harbor. Some methods that use 3D laser scanner sensor or laser range finder to obtain the information in ALS are used presently. But these have some defects respectively in economical efficiency and performance. Therefore, to recover these defects, we propose a method acquiring the information for the movement of a spreader and skew using CCD camera for image processing and laser range finder

  • PDF

Dynamic Manipulation of a Virtual Object in Marker-less AR system Based on Both Human Hands

  • Chun, Jun-Chul;Lee, Byung-Sung
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.4 no.4
    • /
    • pp.618-632
    • /
    • 2010
  • This paper presents a novel approach to control the augmented reality (AR) objects robustly in a marker-less AR system by fingertip tracking and hand pattern recognition. It is known that one of the promising ways to develop a marker-less AR system is using human's body such as hand or face for replacing traditional fiducial markers. This paper introduces a real-time method to manipulate the overlaid virtual objects dynamically in a marker-less AR system using both hands with a single camera. The left bare hand is considered as a virtual marker in the marker-less AR system and the right hand is used as a hand mouse. To build the marker-less system, we utilize a skin-color model for hand shape detection and curvature-based fingertip detection from an input video image. Using the detected fingertips the camera pose are estimated to overlay virtual objects on the hand coordinate system. In order to manipulate the virtual objects rendered on the marker-less AR system dynamically, a vision-based hand control interface, which exploits the fingertip tracking for the movement of the objects and pattern matching for the hand command initiation, is developed. From the experiments, we can prove that the proposed and developed system can control the objects dynamically in a convenient fashion.

High Accuracy Skeleton Estimation using 3D Volumetric Model based on RGB-D

  • Kim, Kyung-Jin;Park, Byung-Seo;Kang, Ji-Won;Kim, Jin-Kyum;Kim, Woo-Suk;Kim, Dong-Wook;Seo, Young-Ho
    • Journal of Broadcast Engineering
    • /
    • v.25 no.7
    • /
    • pp.1095-1106
    • /
    • 2020
  • In this paper, we propose an algorithm that extracts a high-precision 3D skeleton using a model generated using a distributed RGB-D camera. When information about a 3D model is extracted through a distributed RGB-D camera, if the information of the 3D model is used, a skeleton with higher precision can be obtained. In this paper, in order to improve the precision of the 2D skeleton, we find the conditions to obtain the 2D skeleton well using the PCA. Through this, high-quality 2D skeletons are obtained, and high-precision 3D skeletons are extracted by combining the information of the 2D skeletons. Even though this process goes through, the generated skeleton may have errors, so we propose an algorithm that removes these errors by using the information of the 3D model. We were able to extract very high accuracy skeletons using the proposed method.

Semi-automatic 3D Building Reconstruction from Uncalibrated Images (비교정 영상에서의 반자동 3차원 건물 모델링)

  • Jang, Kyung-Ho;Jang, Jae-Seok;Lee, Seok-Jun;Jung, Soon-Ki
    • Journal of Korea Multimedia Society
    • /
    • v.12 no.9
    • /
    • pp.1217-1232
    • /
    • 2009
  • In this paper, we propose a semi-automatic 3D building reconstruction method using uncalibrated images which includes the facade of target building. First, we extract feature points in all images and find corresponding points between each pair of images. Second, we extract lines on each image and estimate the vanishing points. Extracted lines are grouped with respect to their corresponding vanishing points. The adjacency graph is used to organize the image sequence based on the number of corresponding points between image pairs and camera calibration is performed. The initial solid model can be generated by some user interactions using grouped lines and camera pose information. From initial solid model, a detailed building model is reconstructed by a combination of predefined basic Euler operators on half-edge data structure. Automatically computed geometric information is visualized to help user's interaction during the detail modeling process. The proposed system allow the user to get a 3D building model with less user interaction by augmenting various automatically generated geometric information.

  • PDF

Automatic Person Identification using Multiple Cues

  • Swangpol, Danuwat;Chalidabhongse, Thanarat
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1202-1205
    • /
    • 2005
  • This paper describes a method for vision-based person identification that can detect, track, and recognize person from video using multiple cues: height and dressing colors. The method does not require constrained target's pose or fully frontal face image to identify the person. First, the system, which is connected to a pan-tilt-zoom camera, detects target using motion detection and human cardboard model. The system keeps tracking the moving target while it is trying to identify whether it is a human and identify who it is among the registered persons in the database. To segment the moving target from the background scene, we employ a version of background subtraction technique and some spatial filtering. Once the target is segmented, we then align the target with the generic human cardboard model to verify whether the detected target is a human. If the target is identified as a human, the card board model is also used to segment the body parts to obtain some salient features such as head, torso, and legs. The whole body silhouette is also analyzed to obtain the target's shape information such as height and slimness. We then use these multiple cues (at present, we uses shirt color, trousers color, and body height) to recognize the target using a supervised self-organization process. We preliminary tested the system on a set of 5 subjects with multiple clothes. The recognition rate is 100% if the person is wearing the clothes that were learned before. In case a person wears new dresses the system fail to identify. This means height is not enough to classify persons. We plan to extend the work by adding more cues such as skin color, and face recognition by utilizing the zoom capability of the camera to obtain high resolution view of face; then, evaluate the system with more subjects.

  • PDF

Fast Structure Recovery and Integration using Improved Scaled Orthographic Factorization (개선된 직교분해기법을 사용한 빠른 구조 복원 및 융합)

  • Park, Jong-Seung;Yoon, Jong-Hyun
    • Journal of Korea Multimedia Society
    • /
    • v.10 no.3
    • /
    • pp.303-315
    • /
    • 2007
  • This paper proposes a 3D structure recovery and registration method that uses four or more common points. For each frame of a given video, a partial structure is recovered using tracked points. The 3D coordinates, camera positions and camera directions are computed at once by our improved scaled orthographic factorization method. The partially recovered point sets are parts of a whole model. A registration of point sets makes the complete shape. The recovered subsets are integrated by transforming each coordinate system of the local point subset into a common basis coordinate system. The process of shape recovery and integration is performed uniformly and linearly without any nonlinear iterative process and without loss of accuracy. The execution time for the integration is significantly reduced relative to the conventional ICP method. Due to the fast recovery and registration framework, our shape recovery scheme is applicable to various interactive video applications. The processing time per frame is under 0.01 seconds in most cases and the integration error is under 0.1mm on average.

  • PDF

Fixed and Moving Automatic FOD Detection Test using Radar and EO Camera (소형 Radar와 EO 카메라를 이용한 고정형 및 이동형 FOD 자동탐지 시험)

  • Kim, Young-Bin;Kim, Sung-Hee;Park, Myung-Kyu;Park, Kwang-Gun;Kim, Min-su;Hong, Gyo-Young
    • Journal of Advanced Navigation Technology
    • /
    • v.24 no.6
    • /
    • pp.479-484
    • /
    • 2020
  • Foreign object debris (FOD) is a generic term for all substances that may pose a threat to aircraft operations on a runway. In the past, FOD detection and collection methods using human resources were very inefficient in terms of efficiency and economics, so it is essential to develop an unmanned FOD detection system suitable for domestic use. In this paper, the fixed FOD automatic detection system and mobile FOD automatic detection system using EO camera and radar were studied and developed at the Taean airfield of Hanseo University, and fixed and mobile method were operated to confirm that automatic FOD detection in the runway of the airfield is possible regardless of illumination and weather conditions.

Unsupervised Monocular Depth Estimation Using Self-Attention for Autonomous Driving (자율주행을 위한 Self-Attention 기반 비지도 단안 카메라 영상 깊이 추정)

  • Seung-Jun Hwang;Sung-Jun Park;Joong-Hwan Baek
    • Journal of Advanced Navigation Technology
    • /
    • v.27 no.2
    • /
    • pp.182-189
    • /
    • 2023
  • Depth estimation is a key technology in 3D map generation for autonomous driving of vehicles, robots, and drones. The existing sensor-based method has high accuracy but is expensive and has low resolution, while the camera-based method is more affordable with higher resolution. In this study, we propose self-attention-based unsupervised monocular depth estimation for UAV camera system. Self-Attention operation is applied to the network to improve the global feature extraction performance. In addition, we reduce the weight size of the self-attention operation for a low computational amount. The estimated depth and camera pose are transformed into point cloud. The point cloud is mapped into 3D map using the occupancy grid of Octree structure. The proposed network is evaluated using synthesized images and depth sequences from the Mid-Air dataset. Our network demonstrates a 7.69% reduction in error compared to prior studies.

Fast Structure Recovery and Integration using Scaled Orthographic Factorization (개선된 직교분해기법을 사용한 구조의 빠른 복원 및 융합)

  • Yoon, Jong-Hyun;Park, Jong-Seung;Lee, Sang-Rak;Noh, Sung-Ryul
    • 한국HCI학회:학술대회논문집
    • /
    • 2006.02a
    • /
    • pp.486-492
    • /
    • 2006
  • 본 논문에서는 비디오에서의 특징점 추적을 통해 얻은 2D 좌표를 이용한 3D 구조를 추정하는 방법과 네 점 이상의 공통점을 이용한 융합 방법을 제안한다. 영상의 각 프레임에서 공통되는 특징점을 이용하여 형상을 추정한다. 영상의 각 프레임에 대한 특징점의 추적은 Lucas-Kanade 방법을 사용하였다. 3D 좌표 추정 방법으로 개선된 직교분해기법을 사용하였다. 개선된 직교분해기법에서는 3D 좌표를 복원함과 동시에 카메라의 위치와 방향을 계산할 수 있다. 복원된 부분 데이터들은 전체를 이루는 일부분이므로, 융합을 통해 완성된 모습을 만들 수 있다. 복원된 부분 데이터들의 서로 다른 좌표계를 기준 좌표계로 변환함으로써 융합할 수 있다. 융합은 카메라의 모션에 해당하는 카메라의 위치와 방향에 의존된다. 융합 과정은 모두 선형으로 평균 0.5초 이하의 수행 속도를 보이며 융합의 오차는 평균 0.1cm 이하의 오차를 보였다.

  • PDF

3D object recognition using the CAD model and stereo vision

  • Kim, Sung-Il;Choi, Sung-Jun;Won, Sang-Chul
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.669-672
    • /
    • 2003
  • 3D object recognition is difficult but important in computer vision. The important thing is to understand about the relationship between a geometric structure in three dimensions and its image projection. Most 3D recognition systems construct models either manually or by training the pose and orientation of the objects. But both approaches are not satisfactory. In this paper, we focus on a commercial CAD model as a third type of model building for vision. The models are expressed in Initial Graphics Exchanges Specification(IGES) output and reconstructed in a pinhole camera coordinate.

  • PDF