• Title/Summary/Keyword: camera vision

Search Result 1,386, Processing Time 0.03 seconds

A study on the Vision Inspection System for Injection Molding Products (사출제품의 영상검사 시스템 개발에 관한 연구)

  • Shin, Jae-Heung;Kim, Hong-Ryul;Lee, Sang-Cheol;Moon, Sung-Chang
    • Proceedings of the KIEE Conference
    • /
    • 2007.11c
    • /
    • pp.112-116
    • /
    • 2007
  • If any of the set parameters such as the environment temperature, mold temperature are not maintained at a consistent level, the fail rate of injection molding products is increased. The price of the injection molding machine is very high, so in order to maximize the utilization of the machine that is required the production of a number of different products with minimum fail rate using a single machine. To prevent the defect products by an inspection process with perfect quality is very important to minimizing production of defect products in the molding process. Vision inspection systems are widely utilized in various manufacturing industries for quality assurance purposes. The vision inspection system consists of CCD camera and lighting system to capture the image of the subject of inspection, an image comparison algorithm using to determine the pass/fail of the products, and mechanical devices for the operation of the whole system. This research focuses on the development of the vision inspection system to process the inspection of an automobile parts. We developed a mechanical devices for the inspection of the injection molding products and an image comparison algorithm to determine the pass/fail result of the inspection based on the molding image and the accepted product image.

  • PDF

Robust Gaze-Fixing of an Active Vision System under Variation of System Parameters (시스템 파라미터의 변동 하에서도 강건한 능동적인 비전의 시선 고정)

  • Han, Youngmo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.3
    • /
    • pp.195-200
    • /
    • 2012
  • To steer a camera is done based on system parameters of the vision system. However, the system parameters when they are used might be different from those when they were measured. As one method to compensate for this problem, this research proposes a gaze-steering method based on LMI(Linear Matrix Inequality) that is robust to variations in the system parameters of the vision system. Simulation results show that the proposed method produces less gaze-tracking error than a contemporary linear method and more stable gaze-tracking error than a contemporary nonlinear method. Moreover, the proposed method is fast enough for realtime processing.

Design and Fabrication of Multi-rotor system for Vision based Autonomous Landing (영상 기반 자동 착륙용 멀티로터 시스템 설계 및 개발)

  • Kim, Gyou-Beom;Song, Seung-Hwa;Yoon, Kwang-Joon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.12 no.6
    • /
    • pp.141-146
    • /
    • 2012
  • This paper introduces development of multi-rotor system and vision based autonomous landing system. Multi-rotor platform is modeled by rigid body motion with Newton Euler concept. Also Multi-rotor platform is simulated and tuned by LQR control algorithm. Vision based Autonomous Landing system uses a single camera that is mounted Multi-rotor system. Augmented reality algorithm is used as marker detection algorithm and autonomous landing code is test with GCS for the precision landing.

Development of Multi-Laser Vision System For 3D Surface Scanning (3 차원 곡면 데이터 획득을 위한 멀티 레이져 비젼 시스템 개발)

  • Lee, J.H.;Kwon, K.Y.;Lee, H.C.;Doe, Y.C.;Choi, D.J.;Park, J.H.;Kim, D.K.;Park, Y.J.
    • Proceedings of the KSME Conference
    • /
    • 2008.11a
    • /
    • pp.768-772
    • /
    • 2008
  • Various scanning systems have been studied in many industrial areas to acquire a range data or to reconstruct an explicit 3D model. Currently optical technology has been used widely by virtue of noncontactness and high-accuracy. In this paper, we describe a 3D laser scanning system developped to reconstruct the 3D surface of a large-scale object such as a curved-plate of ship-hull. Our scanning system comprises of 4ch-parallel laser vision modules using a triangulation technique. For multi laser vision, calibration method based on least square technique is applied. In global scanning, an effective method without solving difficulty of matching problem among the scanning results of each camera is presented. Also minimal image processing algorithm and robot-based calibration technique are applied. A prototype had been implemented for testing.

  • PDF

Simultaneous Tracking of Multiple Construction Workers Using Stereo-Vision (다수의 건설인력 위치 추적을 위한 스테레오 비전의 활용)

  • Lee, Yong-Ju;Park, Man-Woo
    • Journal of KIBIM
    • /
    • v.7 no.1
    • /
    • pp.45-53
    • /
    • 2017
  • Continuous research efforts have been made on acquiring location data on construction sites. As a result, GPS and RFID are increasingly employed on the site to track the location of equipment and materials. However, these systems are based on radio frequency technologies which require attaching tags on every target entity. Implementing the systems incurs time and costs for attaching/detaching/managing the tags or sensors. For this reason, efforts are currently being made to track construction entities using only cameras. Vision-based 3D tracking has been presented in a previous research work in which the location of construction manpower, vehicle, and materials were successfully tracked. However, the proposed system is still in its infancy and yet to be implemented on practical applications for two reasons. First, it does not involve entity matching across two views, and thus cannot be used for tracking multiple entities, simultaneously. Second, the use of a checker board in the camera calibration process entails a focus-related problem when the baseline is long and the target entities are located far from the cameras. This paper proposes a vision-based method to track multiple workers simultaneously. An entity matching procedure is added to acquire the matching pairs of the same entities across two views which is necessary for tracking multiple entities. Also, the proposed method simplified the calibration process by avoiding the use of a checkerboard, making it more adequate to the realistic deployment on construction sites.

Development of a Pig's Weight Estimating System Using Computer Vision (컴퓨터 시각을 이용한 돼지 무게 예측시스템의 개발)

  • 엄천일;정종훈
    • Journal of Biosystems Engineering
    • /
    • v.29 no.3
    • /
    • pp.275-280
    • /
    • 2004
  • The main objective of this study was to develop and evaluate a model for estimating pigs weight using computer vision for improving the management in Korean swine farms in Korea. This research was carried out in two steps: 1) to find a model that relates the projection area with the weight of a pig; 2) to implement the model in a computer vision system mainly consisted of a monochrome CCD camera, a frame grabber and a computer system for estimating the weight of pigs in a non-contact, real-time manner. The model was developed under an important assumption there were no observable genetic differences among the pigs. The main results were: 1) The relationship between the projection area and the weight of pigs was W = 0.0569 ${\times}$ A - 32.585($R^2$ = 0.953), where W is the weight in kg; A is the projection area of a pig in $\textrm{cm}^2$; 2) The model could estimate the weight of pigs with an error less than 3.5%.

A Parallel Implementation of Multiple Non-overlapping Cameras for Robot Pose Estimation

  • Ragab, Mohammad Ehab;Elkabbany, Ghada Farouk
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.11
    • /
    • pp.4103-4117
    • /
    • 2014
  • Image processing and computer vision algorithms are gaining larger concern in a variety of application areas such as robotics and man-machine interaction. Vision allows the development of flexible, intelligent, and less intrusive approaches than most of the other sensor systems. In this work, we determine the location and orientation of a mobile robot which is crucial for performing its tasks. In order to be able to operate in real time there is a need to speed up different vision routines. Therefore, we present and evaluate a method for introducing parallelism into the multiple non-overlapping camera pose estimation algorithm proposed in [1]. In this algorithm the problem has been solved in real time using multiple non-overlapping cameras and the Extended Kalman Filter (EKF). Four cameras arranged in two back-to-back pairs are put on the platform of a moving robot. An important benefit of using multiple cameras for robot pose estimation is the capability of resolving vision uncertainties such as the bas-relief ambiguity. The proposed method is based on algorithmic skeletons for low, medium and high levels of parallelization. The analysis shows that the use of a multiprocessor system enhances the system performance by about 87%. In addition, the proposed design is scalable, which is necaccery in this application where the number of features changes repeatedly.

Implementation of Real-Time Post-Processing for High-Quality Stereo Vision

  • Choi, Seungmin;Jeong, Jae-Chan;Chang, Jiho;Shin, Hochul;Lim, Eul-Gyoon;Cho, Jae Il;Hwang, Daehwan
    • ETRI Journal
    • /
    • v.37 no.4
    • /
    • pp.752-765
    • /
    • 2015
  • We propose a novel post-processing algorithm and its very-large-scale integration architecture that simultaneously uses the passive and active stereo vision information to improve the reliability of the three-dimensional disparity in a hybrid stereo vision system. The proposed architecture consists of four steps - left-right consistency checking, semi-2D hole filling, a tiny adaptive variance checking, and a 2D weighted median filter. The experimental results show that the error rate of the proposed algorithm (5.77%) is less than that of a raw disparity (10.12%) for a real-world camera image having a $1,280{\times}720$ resolution and maximum disparity of 256. Moreover, for the famous Middlebury stereo image sets, the proposed algorithm's error rate (8.30%) is also less than that of the raw disparity (13.7%). The proposed architecture is implemented on a single commercial field-programmable gate array using only 13.01% of slice resources, which achieves a rate of 60 fps for $1,280{\times}720$ stereo images with a disparity range of 256.

Development of Auto Sorting System for T Type Welding nut using A Vision Inspector (비전 검사기를 활용한 T형 용접너트 자동 선별시스템 개발)

  • Song, Han-Lim;Hur, Tae-Won
    • 전자공학회논문지 IE
    • /
    • v.48 no.1
    • /
    • pp.16-24
    • /
    • 2011
  • In this paper, we developed a auto sorting system for T type welding nut using a vision inspector. We used edge and thread detection with histogram of image which is captured by machine vision camera. We also used a binary morphology operation for a detection of spot. As a result we performed numeric inspection of 0.1mm accuracy. This is impossible in old sorting system and inspector with naked eye. Also, we reduced the manufacturing unit cost to 25% and improved a production efficiency to 330%.

Detection of Surface Cracks in Eggshell by Machine Vision and Artificial Neural Network (기계 시각과 인공 신경망을 이용한 파란의 판별)

  • 이수환;조한근;최완규
    • Journal of Biosystems Engineering
    • /
    • v.25 no.5
    • /
    • pp.409-414
    • /
    • 2000
  • A machine vision system was built to obtain single stationary image from an egg. This system includes a CCD camera, an image processing board and a lighting system. A computer program was written to acquire, enhance and get histogram from an image. To minimize the evaluation time, the artificial neural network with the histogram of the image was used for eggshell evaluation. Various artificial neural networks with different parameters were trained and tested. The best network(64-50-1 and 128-10-1) showed an accuracy of 87.5% in evaluating eggshell. The comparison test for the elapsed processing time per an egg spent by this method(image processing and artificial neural network) and by the processing time per an egg spent by this method(image processing and artificial neural network) and by the previous method(image processing only) revealed that it was reduced to about a half(5.5s from 10.6s) in case of cracked eggs and was reduced to about one-fifth(5.5s from 21.1s) in case of normal eggs. This indicates that a fast eggshell evaluation system can be developed by using machine vision and artificial neural network.

  • PDF