• Title/Summary/Keyword: Plane Homography

Search Result 18, Processing Time 0.02 seconds

Ground Plane Detection Using Homography Matrix (호모그래피행렬을 이용한 노면검출)

  • Lee, Ki-Yong;Lee, Joon-Woong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.17 no.10
    • /
    • pp.983-988
    • /
    • 2011
  • This paper presents a robust method for ground plane detection in vision-based applications based on a monocular sequence of images with a non-stationary camera. The proposed method, which is based on the reliable estimation of the homography between two frames taken from the sequence, aims at designing a practical system to detect road surface from traffic scenes. The homography is computed using a feature matching approach, which often gives rise to inaccurate matches or undesirable matches from out of the ground plane. Hence, the proposed homography estimation minimizes the effects from erroneous feature matching by the evaluation of the difference between the predicted and the observed matrices. The method is successfully demonstrated for the detection of road surface performed on experiments to fill an information void area taken place from geometric transformation applied to captured images by an in-vehicle camera system.

Multiple Camera-Based Correspondence of Ground Foot for Human Motion Tracking (사람의 움직임 추적을 위한 다중 카메라 기반의 지면 위 발의 대응)

  • Seo, Dong-Wook;Chae, Hyun-Uk;Jo, Kang-Hyun
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.14 no.8
    • /
    • pp.848-855
    • /
    • 2008
  • In this paper, we describe correspondence among multiple images taken by multiple cameras. The correspondence among multiple views is an interesting problem which often appears in the application like visual surveillance or gesture recognition system. We use the principal axis and the ground plane homography to estimate foot of human. The principal axis belongs to the subtracted silhouette-based region of human using subtraction of the predetermined multiple background models with current image which includes moving person. For the calculation of the ground plane homography, we use landmarks on the ground plane in 3D space. Thus the ground plane homography means the relation of two common points in different views. In the normal human being, the foot of human has an exactly same position in the 3D space and we represent it to the intersection in this paper. The intersection occurs when the principal axis in an image crosses to the transformed ground plane from other image. However the positions of the intersection are different depend on camera views. Therefore we construct the correspondence that means the relationship between the intersection in current image and the transformed intersection from other image by homography. Those correspondences should confirm within a short distance measuring in the top viewed plane. Thus, we track a person by these corresponding points on the ground plane. Experimental result shows the accuracy of the proposed algorithm has almost 90% of detecting person for tracking based on correspondence of intersections.

Detection of the co-planar feature points in the three dimensional space (3차원 공간에서 동일 평면 상에 존재하는 특징점 검출 기법)

  • Seok-Han Lee
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.16 no.6
    • /
    • pp.499-508
    • /
    • 2023
  • In this paper, we propose a technique to estimate the coordinates of feature points existing on a 2D planar object in the three dimensional space. The proposed method detects multiple 3D features from the image, and excludes those which are not located on the plane. The proposed technique estimates the planar homography between the planar object in the 3D space and the camera image plane, and computes back-projection error of each feature point on the planar object. Then any feature points which have large error is considered as off-plane points and are excluded from the feature estimation phase. The proposed method is archived on the basis of the planar homography without any additional sensors or optimization algorithms. In the expretiments, it was confirmed that the speed of the proposed method is more than 40 frames per second. In addition, compared to the RGB-D camera, there was no significant difference in processing speed, and it was verified that the frame rate was unaffected even in the situation that the number of detected feature points continuously increased.

Tele-presence System using Homography-based Camera Tracking Method (호모그래피기반의 카메라 추적기술을 이용한 텔레프레즌스 시스템)

  • Kim, Tae-Hyub;Choi, Yoon-Seok;Nam, Bo-Dam;Hong, Hyun-Ki
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.49 no.3
    • /
    • pp.27-33
    • /
    • 2012
  • Tele-presence and tele-operation techniques are used to build up an immersive scene and control environment for the distant user. This paper presents a novel tele-presence system using the camera tracking based on planar homography. In the first step, the user wears the HMD(head mounted display) with the camera and his/her head motion is estimated. From the panoramic image by the omni-directional camera mounted on the mobile robot, a viewing image by the user is generated and displayed through HMD. The homography of 3D plane with markers is used to obtain the head motion of the user. For the performance evaluation, the camera tracking results by ARToolkit and the homography based method are compared with the really measured positions of the camera.

Performing Missions of a Minicar Using a Single Camera (단안 카메라를 이용한 소형 자동차의 임무 수행)

  • Kim, Jin-Woo;Ha, Jong-Eun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.12 no.1
    • /
    • pp.123-128
    • /
    • 2017
  • This paper deals with performing missions through autonomous navigation using camera and other sensors. Extracting pose of the car is necessary to navigate safely within the given road. Homography is used to find it. Color image is converted into grey image and thresholding and edge is used to find control points. Two control ponits are converted into world coordinates using homography to find the angle and position of the car. Color is used to find traffic signal. It was confirmed that the given tasks performed well through experiments.

A New Shape-Based Object Category Recognition Technique using Affine Category Shape Model (Affine Category Shape Model을 이용한 형태 기반 범주 물체 인식 기법)

  • Kim, Dong-Hwan;Choi, Yu-Kyung;Park, Sung-Kee
    • The Journal of Korea Robotics Society
    • /
    • v.4 no.3
    • /
    • pp.185-191
    • /
    • 2009
  • This paper presents a new shape-based algorithm using affine category shape model for object category recognition and model learning. Affine category shape model is a graph of interconnected nodes whose geometric interactions are modeled using pairwise potentials. In its learning phase, it can efficiently handle large pose variations of objects in training images by estimating 2-D homography transformation between the model and the training images. Since the pairwise potentials are defined on only relative geometric relationship betweenfeatures, the proposed matching algorithm is translation and in-plane rotation invariant and robust to affine transformation. We apply spectral matching algorithm to find feature correspondences, which are then used as initial correspondences for RANSAC algorithm. The 2-D homography transformation and the inlier correspondences which are consistent with this estimate can be efficiently estimated through RANSAC, and new correspondences also can be detected by using the estimated 2-D homography transformation. Experimental results on object category database show that the proposed algorithm is robust to pose variation of objects and provides good recognition performance.

  • PDF

Analysis of Rotational Motion of Skid Steering Mobile Robot using Marker and Camera (마커와 카메라를 이용한 스키드 구동 이동 로봇의 회전 운동 분석)

  • Ha, Jong-Eun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.11 no.2
    • /
    • pp.185-190
    • /
    • 2016
  • This paper deals with analysis of the characteristics of mobile robot's motion by automatic detection of markers on a robot using a camera. Analysis of motion behaviors according to parameters is important in developing control algorithm for robot operation or autonomous navigation. For this purpose, we use four chessboard patterns on the robot. Their location on the robot is adjusted to be on single plane. Homography is used to compute the actual amount of movement of the robot. Presented method is tested using P3-AT robot and it gives reliable results.

Passive Ranging Based on Planar Homography in a Monocular Vision System

  • Wu, Xin-mei;Guan, Fang-li;Xu, Ai-jun
    • Journal of Information Processing Systems
    • /
    • v.16 no.1
    • /
    • pp.155-170
    • /
    • 2020
  • Passive ranging is a critical part of machine vision measurement. Most of passive ranging methods based on machine vision use binocular technology which need strict hardware conditions and lack of universality. To measure the distance of an object placed on horizontal plane, we present a passive ranging method based on monocular vision system by smartphone. Experimental results show that given the same abscissas, the ordinatesis of the image points linearly related to their actual imaging angles. According to this principle, we first establish a depth extraction model by assuming a linear function and substituting the actual imaging angles and ordinates of the special conjugate points into the linear function. The vertical distance of the target object to the optical axis is then calculated according to imaging principle of camera, and the passive ranging can be derived by depth and vertical distance to the optical axis of target object. Experimental results show that ranging by this method has a higher accuracy compare with others based on binocular vision system. The mean relative error of the depth measurement is 0.937% when the distance is within 3 m. When it is 3-10 m, the mean relative error is 1.71%. Compared with other methods based on monocular vision system, the method does not need to calibrate before ranging and avoids the error caused by data fitting.

Object Width Measurement System Using Light Sectioning Method (광절단법을 이용한 물체 크기 측정 시스템)

  • Lee, Byeong-Ju;Kang, Hyun-Soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.18 no.3
    • /
    • pp.697-705
    • /
    • 2014
  • This paper presents a vision based object width measurement method and its application where the light sectioning method is employed. The target object for measurement is a tread, which is the most outside component of an automobile tire. The entire system applying the measurement method consists of two processes, i.e. a calibration process and a detection process. The calibration process is to identify the relationships between a camera plane and a laser plane, and to estimate a camera lens distortion parameters. As the process requires a test pattern, namely a jig, which is elaborately manufactured. In the detection process, first of all, the region that a laser light illuminates is extracted by applying an adaptive thresholding technique where the distribution of the pixel brightness is considered to decide the optimal threshold. Then, a thinning algorithm is applied to the region so that the ends and the shoulders of a tread are detected. Finally, the tread width and the shoulder width are computed using the homography and the distortion coefficients obtained by the calibration process.

Target-free vision-based approach for vibration measurement and damage identification of truss bridges

  • Dong Tan;Zhenghao Ding;Jun Li;Hong Hao
    • Smart Structures and Systems
    • /
    • v.31 no.4
    • /
    • pp.421-436
    • /
    • 2023
  • This paper presents a vibration displacement measurement and damage identification method for a space truss structure from its vibration videos. Features from Accelerated Segment Test (FAST) algorithm is combined with adaptive threshold strategy to detect the feature points of high quality within the Region of Interest (ROI), around each node of the truss structure. Then these points are tracked by Kanade-Lucas-Tomasi (KLT) algorithm along the video frame sequences to obtain the vibration displacement time histories. For some cases with the image plane not parallel to the truss structural plane, the scale factors cannot be applied directly. Therefore, these videos are processed with homography transformation. After scale factor adaptation, tracking results are expressed in physical units and compared with ground truth data. The main operational frequencies and the corresponding mode shapes are identified by using Subspace Stochastic Identification (SSI) from the obtained vibration displacement responses and compared with ground truth data. Structural damages are quantified by elemental stiffness reductions. A Bayesian inference-based objective function is constructed based on natural frequencies to identify the damage by model updating. The Success-History based Adaptive Differential Evolution with Linear Population Size Reduction (L-SHADE) is applied to minimise the objective function by tuning the damage parameter of each element. The locations and severities of damage in each case are then identified. The accuracy and effectiveness are verified by comparison of the identified results with the ground truth data.