• Title/Summary/Keyword: line camera

Search Result 820, Processing Time 0.025 seconds

Localization for Mobile Robot Using Vertical Line Features (수직선 특징을 이용한 이동 로봇의 자기 위치 추정)

  • 강창훈;안현식
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.9 no.11
    • /
    • pp.937-942
    • /
    • 2003
  • We present a self-localization method for mobile robots using vertical line features of indoor environment. When a 2D map including feature points and color information is given, a mobile robot moves to the destination, and acquires images from the surroundings having vertical line edges by one camera. From the image, vertical line edges are detected, and pattern vectors meaning averaged color values of the left and right regions of the each line are computed by using the properties of the line and a region growing method. The pattern vectors are matched with the feature points of the map by comparing the color information and the geometrical relationship. From the perspective transformation and rigid transformation of the corresponded points, nonlinear equations are derived. Localization is carried out from solving the equations by using Newton's method. Experimental results show that the proposed method using mono view is simple and applicable to indoor environment.

Text Line Segmentation of Handwritten Documents by Area Mapping

  • Boragule, Abhijeet;Lee, GueeSang
    • Smart Media Journal
    • /
    • v.4 no.3
    • /
    • pp.44-49
    • /
    • 2015
  • Text line segmentation is a preprocessing step in OCR, which can significantly influence the accuracy of document analysis applications. This paper proposes a novel methodology for the text line segmentation of handwritten documents. First, the average width of the connected components is used to form a 1-D Gaussian kernel and a smoothing operation is then applied to the input binary image. The adaptive binarization of the smoothed image forms the final text lines. In this work, the segmentation method involves two stages: firstly, the large connected components are labelled as a unique text line using text line area mapping. Secondly, the final refinement of the segmentation is performed using the Euclidean distance between the text line and small connected components. The group of uniquely labelled text candidates achieves promising segmentation results. The proposed approach works well on Korean and English language handwritten documents captured using a camera.

Omnidirectional Camera Motion Estimation Using Projected Contours (사영 컨투어를 이용한 전방향 카메라의 움직임 추정 방법)

  • Hwang, Yong-Ho;Lee, Jae-Man;Hong, Hyun-Ki
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.5
    • /
    • pp.35-44
    • /
    • 2007
  • Since the omnidirectional camera system with a very large field of view could take many information about environment scene from few images, various researches for calibration and 3D reconstruction using omnidirectional image have been presented actively. Most of line segments of man-made objects we projected to the contours by using the omnidirectional camera model. Therefore, the corresponding contours among images sequences would be useful for computing the camera transformations including rotation and translation. This paper presents a novel two step minimization method to estimate the extrinsic parameters of the camera from the corresponding contours. In the first step, coarse camera parameters are estimated by minimizing an angular error function between epipolar planes and back-projected vectors from each corresponding point. Then we can compute the final parameters minimizing a distance error of the projected contours and the actual contours. Simulation results on the synthetic and real images demonstrated that our algorithm can achieve precise contour matching and camera motion estimation.

Experimental Setup for Autonomous Navigation of Robotic Vehicle for University Campus (대학 캠퍼스용 로봇차량의 자율주행을 위한 실험환경 구축)

  • Cho, Sung Taek;Park, Young Jun;Jung, Seul
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.26 no.2
    • /
    • pp.105-112
    • /
    • 2016
  • This paper presents the experimental setup for autonomous navigation of a robotic vehicle for touring university campus. The robotic vehicle is developed for navigation of specific areas such as university campus or play parks. The robotic vehicle can carry two passengers to travel short distances. For the robotic vehicle to navigate autonomously the specific distance from the main gate to the administrative building in the university, the experimental setup for SLAM is presented. As an initial step, a simple method of following the line detected by a single camera is implemented for the partial area. The central line on the pavement colored with two kinds, red and yellow, is detected by image processing, and the robotic vehicle is commanded to follow the line. Experimental studies are conducted to demonstrate the performance of navigation as a possible touring vehicle.

011-line Visual Feedback Control of Industrial Robot Manipulator (산업용 로봇 매니퓰레이터의 오프라인 영상피드백 제어)

  • 신행봉;정동연;김용태;이종두;이강두
    • Proceedings of the Korean Society of Machine Tool Engineers Conference
    • /
    • 2002.04a
    • /
    • pp.567-572
    • /
    • 2002
  • The equipment of industrial robot in manufacturing and assembly lines has rapidly increased. In order to achieve high productivity and flexibility, it becomes very important to develop the visual feedback control system with Off-Line Programming System(OLPS ). We can save much efforts and time in adjusting robots to newly defined workcells by using Off-Line Programming System. A proposed visual calibration scheme is based on position-based visual feedback. The visual calibration system is composed of a personal computer, an image processing board, a video monitor, and one camera. The calibration program firstly generates predicted images of objects in an assumed end-effector position. The process to generate predicted images consists of projection to screen-coordinates, visible range test, and construction of simple silhouette figures. Then, camera images acquired are compared with predicted ones for updating position and orientation data. Computation of error is very simple because the scheme is based on perspective projection, which can be also expanded to experimental results. Computation time can be extremely reduced because the proposed method does not require the precise calculation of tree-dimensional object data and image Jacobian.

  • PDF

A Yarn Process Inspection System Using Image Processing (영상처리를 이용한 원사공정 검사시스템)

  • Lim, Chang-Yong;Shin, Dongwon;Yoon, Jang-Kyu
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.30 no.5
    • /
    • pp.513-519
    • /
    • 2013
  • Line scan camera has been widely used in the area of inspection of glass, film, fabric, iron, PCB and etc. due to the high resolution and the high speed. We developed the line scan based vision system to inspect tangled and cut-off status of yarn in the manufacturing process. The original image is binarized with a proper threshold, and the gap distances in the yarn are measured in real time, so finally the status of the process is decided by the maximum value of the gap distance. All procedures are executed in real time by realization of multi-processed threads. By implementation of this system, the error of the yarn in manufacturing process can be precedently monitored and the loss of the yarn is decreased efficiently.

Improved LiDAR-Camera Calibration Using Marker Detection Based on 3D Plane Extraction

  • Yoo, Joong-Sun;Kim, Do-Hyeong;Kim, Gon-Woo
    • Journal of Electrical Engineering and Technology
    • /
    • v.13 no.6
    • /
    • pp.2530-2544
    • /
    • 2018
  • In this paper, we propose an enhanced LiDAR-camera calibration method that extracts the marker plane from 3D point cloud information. In previous work, we estimated the straight line of each board to obtain the vertex. However, the errors in the point information in relation to the z axis were not considered. These errors are caused by the effects of user selection on the board border. Because of the nature of LiDAR, the point information is separated in the horizontal direction, causing the approximated model of the straight line to be erroneous. In the proposed work, we obtain each vertex by estimating a rectangle from a plane rather than obtaining a point from each straight line in order to obtain a vertex more precisely than the previous study. The advantage of using planes is that it is easier to select the area, and the most point information on the board is available. We demonstrated through experiments that the proposed method could be used to obtain more accurate results compared to the performance of the previous method.

Off-line Visual Feedback Control of Robot Manipulator (로봇 매니퓰레이터의 오프라인 영상피드백 제어)

  • 신행봉;정동연;이종두;이강두;한성현
    • Proceedings of the Korean Society of Machine Tool Engineers Conference
    • /
    • 2001.10a
    • /
    • pp.140-145
    • /
    • 2001
  • The equipment of industrial robot in manufacturing and assembly lines has rapidly increased. In order to achieve high productivity and flexibility, it becomes very important to develop the visual feedback control system with Off-Line Programming System(OLPS). We can save much efforts and time in adjusting robots to newly defined workcells by using Off-Line Programming System. A proposed visual calibration scheme is based on position-based visual feedback. The visual calibration system is composed of a personal computer, an image processing board, a video monitor, and one camera. The calibration program firstly generates predicted images of objects in an assumed end-effector position. The process to generate predicted images consists of projection to screen-coordinates, visible range test, and construction of simple silhouette figures. Then, camera images acquired are compared with predicted ones for updating position and orientation data. Computation of error is very simple because the scheme is based on perspective projection, which can be also expanded to experimental results. Computation time can be extremely reduced because the proposed method does not require the precise calculation of tree-dimensional object data and image Jacobian.

  • PDF

Development of Automated Surface Inspection System using the Computer V (컴퓨터 비젼을 이용한 표면결함검사장치 개발)

  • Lee, Jong-Hak;Jung, Jin-Yang
    • Proceedings of the KIEE Conference
    • /
    • 1999.07b
    • /
    • pp.668-670
    • /
    • 1999
  • We have developed a automatic surface inspection system for cold Rolled strips in steel making process for several years. We have experienced the various kinds of surface inspection systems, including linear CCD camera type and the laser type inspection system which was installed in cold rolled strips production lines. But, we did not satisfied with these inspection systems owing to insufficient detection and classification rate, real time processing performance and limited line speed of real production lines. In order to increase detection and computing power, we have used the Dark Field illumination with Infra_Red LED, Bright Field illumination with Xenon Lamp, Parallel Computing Processor with Area typed CCD camera and full software based image processing technique for the ease up_grading and maintenance. In this paper, we introduced the automatic inspection system and real time image processing technique using the Object Detection, Defect Detection, Classification algorithms. As a result of experiment, under the situation of the high speed processed line(max 1000 meter per minute) defect detection is above 90% for all occurred defects in real line, defect name classification rate is about 80% for most frequently occurred 8 defect, and defect grade classification rate is 84% for name classified defect.

  • PDF