• 제목/요약/키워드: Vision-Based Navigation

검색결과 193건 처리시간 0.024초

무인 컨테이너 운반차량의 장애물 인식을 위한 물체의 위치 및 변위 검출에 관한 연구 (A Study on Detection of Object Position and Displacement for Obstacle Recognition of UCT)

  • 이진우;이영진;조현철;손주한;이권순
    • 한국항해항만학회:학술대회논문집
    • /
    • 한국항해항만학회 1999년도 추계학술대회논문집
    • /
    • pp.321-332
    • /
    • 1999
  • It is important to detect objects movement for obstacle recognition and path searching of UCT(unmanned container transporters) with vision sensor. This paper shows the method to draw out objects and to trace the trajectory of the moving object using a CCD camera and it describes the method to recognize the shape of objects by neural network. We can transform pixel points to objects position of the real space using the proposed viewport. This proposed technique is used by the single vision system based on floor map.

  • PDF

드론 스트리밍 영상 이미지 분석을 통한 실시간 산불 탐지 시스템 (Forest Fire Detection System using Drone Streaming Images)

  • Yoosin Kim
    • 한국항행학회논문지
    • /
    • 제27권5호
    • /
    • pp.685-689
    • /
    • 2023
  • The proposed system in the study aims to detect forest fires in real-time stream data received from the drone-camera. Recently, the number of wildfires has been increasing, and also the large scaled wildfires are frequent more and more. In order to prevent forest fire damage, many experiments using the drone camera and vision analysis are actively conducted, however there were many challenges, such as network speed, pre-processing, and model performance, to detect forest fires from real-time streaming data of the flying drone. Therefore, this study applied image data processing works to capture five good image frames for vision analysis from whole streaming data and then developed the object detection model based on YOLO_v2. As the result, the classification model performance of forest fire images reached upto 93% of accuracy, and the field test for the model verification detected the forest fire with about 70% accuracy.

Limit-cycle 항법과 모서리 검출을 기반으로 하는 UGV를 위한 계획 경로 알고리즘 (Path Planning Algorithm for UGVs Based on the Edge Detecting and Limit-cycle Navigation Method)

  • 임윤원;정진수;안진웅;김동한
    • 제어로봇시스템학회논문지
    • /
    • 제17권5호
    • /
    • pp.471-478
    • /
    • 2011
  • This UGV (Unmanned Ground Vehicle) is not only widely used in various practical applications but is also currently being researched in many disciplines. In particular, obstacle avoidance is considered one of the most important technologies in the navigation of an unmanned vehicle. In this paper, we introduce a simple algorithm for path planning in order to reach a destination while avoiding a polygonal-shaped static obstacle. To effectively avoid such an obstacle, a path planned near the obstacle is much shorter than a path planned far from the obstacle, on the condition that both paths guarantee that the robot will not collide with the obstacle. So, to generate a path near the obstacle, we have developed an algorithm that combines an edge detection method and a limit-cycle navigation method. The edge detection method, based on Hough Transform and IR sensors, finds an obstacle's edge, and the limit-cycle navigation method generates a path that is smooth enough to reach a detected obstacle's edge. And we proposed novel algorithm to solve local minima using the virtual wall in the local vision. Finally, we verify performances of the proposed algorithm through simulations and experiments.

광추적기와 내부 비전센서를 이용한 수술도구의 3차원 자세 및 위치 추적 시스템 (3D Orientation and Position Tracking System of Surgical Instrument with Optical Tracker and Internal Vision Sensor)

  • 조영진;오현민;김민영
    • 제어로봇시스템학회논문지
    • /
    • 제22권8호
    • /
    • pp.579-584
    • /
    • 2016
  • When surgical instruments are tracked in an image-guided surgical navigation system, a stereo vision system with high accuracy is generally used, which is called optical tracker. However, this optical tracker has the disadvantage that a line-of-sight between the tracker and surgical instrument must be maintained. Therefore, to complement the disadvantage of optical tracking systems, an internal vision sensor is attached to a surgical instrument in this paper. Monitoring the target marker pattern attached on patient with this vision sensor, this surgical instrument is possible to be tracked even when the line-of-sight of the optical tracker is occluded. To verify the system's effectiveness, a series of basic experiments is carried out. Lastly, an integration experiment is conducted. The experimental results show that rotational error is bounded to max $1.32^{\circ}$ and mean $0.35^{\circ}$, and translation error is in max 1.72mm and mean 0.58mm. Finally, it is confirmed that the proposed tool tracking method using an internal vision sensor is useful and effective to overcome the occlusion problem of the optical tracker.

실내 주행을 위한 3차원 장애물 검출 (Three Dimensional Obstacle Detection for Indoor Navigation)

  • 고복경;우동민
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 1996년도 하계학술대회 논문집 B
    • /
    • pp.1251-1253
    • /
    • 1996
  • The vision processing system for mobile robots requires real time processing and reliability for the purpose of safe navigation. But, general types of vision systems are not appropriate owing to the correspondence problem which correlates the points out of two images. To determine the obstacle area, we use correspondences of line segments between two perspective images sequentially acquired by camera. To simplify the correspondence, the matching of line segments are performed in the navigation space, based on the assumption that mobile robot should be navigated in the flat surface and the motion of mobile robot between two frames should be approximately known.

  • PDF

자율주행차량을 위한 비젼 기반의 횡방향 제어 시스템 개발 (Development of Vision-based Lateral Control System for an Autonomous Navigation Vehicle)

  • 노광현
    • 한국자동차공학회논문집
    • /
    • 제13권4호
    • /
    • pp.19-25
    • /
    • 2005
  • This paper presents a lateral control system for the autonomous navigation vehicle that was developed and tested by Robotics Centre of Ecole des Mines do Paris in France. A robust lane detection algorithm was developed for detecting different types of lane marker in the images taken by a CCD camera mounted on the vehicle. $^{RT}Maps$ that is a software framework far developing vision and data fusion applications, especially in a car was used for implementing lane detection and lateral control. The lateral control has been tested on the urban road in Paris and the demonstration has been shown to the public during IEEE Intelligent Vehicle Symposium 2002. Over 100 people experienced the automatic lateral control. The demo vehicle could run at a speed of 130km1h in the straight road and 50km/h in high curvature road stably.

이동로봇의 자율주행을 위한 다중센서융합기반의 지도작성 및 위치추정 (Map-Building and Position Estimation based on Multi-Sensor Fusion for Mobile Robot Navigation in an Unknown Environment)

  • 진태석;이민중;이장명
    • 제어로봇시스템학회논문지
    • /
    • 제13권5호
    • /
    • pp.434-443
    • /
    • 2007
  • Presently, the exploration of an unknown environment is an important task for thee new generation of mobile service robots and mobile robots are navigated by means of a number of methods, using navigating systems such as the sonar-sensing system or the visual-sensing system. To fully utilize the strengths of both the sonar and visual sensing systems. This paper presents a technique for localization of a mobile robot using fusion data of multi-ultrasonic sensors and vision system. The mobile robot is designed for operating in a well-structured environment that can be represented by planes, edges, comers and cylinders in the view of structural features. In the case of ultrasonic sensors, these features have the range information in the form of the arc of a circle that is generally named as RCD(Region of Constant Depth). Localization is the continual provision of a knowledge of position which is deduced from it's a priori position estimation. The environment of a robot is modeled into a two dimensional grid map. we defines a vision-based environment recognition, phisically-based sonar sensor model and employs an extended Kalman filter to estimate position of the robot. The performance and simplicity of the approach is demonstrated with the results produced by sets of experiments using a mobile robot.

Vision을 이용한 자율주행 로봇의 라인 인식 및 주행방향 결정에 관한 연구 (A Study of Line Recognition and Driving Direction Control On Vision based AGV)

  • 김영숙;김태완;이창구
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 2002년도 하계학술대회 논문집 D
    • /
    • pp.2341-2343
    • /
    • 2002
  • This paper describes a vision-based line recognition and control of driving direction for an AGV(autonomous guided vehicle). As navigation guide, black stripe attached on the corridor is used. Binary image of guide stripe captured by a CCD camera is used. For detect the guideline quickly and extractly, we use for variable thresholding algorithm. this low-cost line-tracking system is efficiently using pc-based real time vision processing. steering control is studied through controller with guide-line angle error. This method is tested via a typical agv with a single camera in laboratory environment.

  • PDF

무인 시스템의 자율 주행을 위한 영상기반 항법기술 동향 (Survey on Visual Navigation Technology for Unmanned Systems)

  • 김현진;서호성;김표진;이충근
    • 한국항행학회논문지
    • /
    • 제19권2호
    • /
    • pp.133-139
    • /
    • 2015
  • 이 논문에서는 영상정보를 기반으로 한 무인 시스템의 자율 항법기술에 대한 동향을 요약한다. 영상기반 항법기술로는 비주얼 서보잉, 비주얼 오도메트리, 영상 기반 SLAM(simultaneous localization and mapping)이 있다. 비주얼 서보잉은 목표 이미지와 현재 이미지 사이의 피쳐 차이로부터 원하는 속도 입력을 계산하여 무인 로봇을 목표 자세로 유도하는 데 사용된다. 비주얼 오도메트리는 무인 시스템이 영상정보를 바탕으로 자신의 이동 궤적을 추정하는 기술로, 기존의 dead-reckoning 방식보다 정확성을 향상시킬 수 있다. 영상 기반 SLAM은 무인 시스템이 영상 정보를 활용하여 미지의 환경에 대한 지도를 구축함과 동시에 자신의 위치를 결정해 나가는 기술로, 정확히 알지 못하는 환경에서 무인차량이나 무인기를 운용하는데 필수적이다. 이러한 기술들이 적용된 해외의 연구 사례들을 살펴봄으로써 영상기반 항법기술의 동향을 파악할 수 있었다.

REPRESENTATION OF NAVIGATION INFORMATION FOR VISUAL CAR NAVIGATION SYSTEM

  • Joo, In-Hak;Lee, Seung-Yong;Cho, Seong-Ik
    • 대한원격탐사학회:학술대회논문집
    • /
    • 대한원격탐사학회 2007년도 Proceedings of ISRS 2007
    • /
    • pp.508-511
    • /
    • 2007
  • Car navigation system is one of the most important applications in telematics. A newest trend of car navigation system is using real video captured by camera equipped on the vehicle, because video can overcome the semantic gap between map and real world. In this paper, we suggest a visual car navigation system that visually represents navigation information or route guidance. It can improve drivers' understanding about real world by capturing real-time video and displaying navigation information overlaid on it. Main services of the visual car navigation system are graphical turn guidance and lane change guidance. We suggest the system architecture that implements the services by integrating conventional route finding and guidance, computer vision functions, and augmented reality display functions. What we designed as a core part of the system is visual navigation controller, which controls other modules and dynamically determines visual representation methods of navigation information according to a determination rule based on current location and driving circumstances. We briefly show the implementation of system.

  • PDF