• Title/Summary/Keyword: Vision sensor

Search Result 832, Processing Time 0.028 seconds

Trends of Sensor-based Intelligent Arc Welding Robot System (센서기반 지능형 아크 용접 로봇 시스템의 동향)

  • Joung, Ji Hoon;Shin, Hyeon-Ho;Song, Young Hoon;Kim, SooJong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.10
    • /
    • pp.1051-1056
    • /
    • 2014
  • In this paper, we introduce an intelligent robotic arc welding system which exploits sensors like as LVS (Laser Vision Sensor), Hall effect sensor, voltmeter and so on. The use of industrial robot is saturated because of its own limitation, and one of the major limitations is that industrial robot cannot recognize the environment. Lately, sensor-based environmental awareness research of the industrial robot is performed actively to overcome such limitation, and it can expand application field and improve productivity. We classify the sensor-based intelligent arc welding robot system by the goal and the sensing data. The goals can be categorized into detection of a welding start point, tracking of a welding line and correction of a torch deformation. The Sensing data can be categorized into welding data (i.e. current, voltage and short circuit detection) and displacement data (i.e. distance, position). This paper covers not only the explanation of the each category but also its advantage and limitation.

Intelligent Shoes for Detecting Blind Falls Using the Internet of Things

  • Ahmad Abusukhon
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.9
    • /
    • pp.2377-2398
    • /
    • 2023
  • In our daily lives, we engage in a variety of tasks that rely on our senses, such as seeing. Blindness is the absence of the sense of vision. According to the World Health Organization, 2.2 billion people worldwide suffer from various forms of vision impairment. Unfortunately, blind people face a variety of indoor and outdoor challenges on a daily basis, limiting their mobility and preventing them from engaging in other activities. Blind people are very vulnerable to a variety of hazards, including falls. Various barriers, such as stairs, can cause a fall. The Internet of Things (IoT) is used to track falls and send a warning message to the blind caretakers. One of the gaps in the previous works is that they were unable to differentiate between falls true and false. Treating false falls as true falls results in many false alarms being sent to the blind caretakers and thus, they may reject the IoT system. As a means of bridging this chasm, this paper proposes an intelligent shoe that is able to precisely distinguish between false and true falls based on three sensors, namely, the load scale sensor, the light sensor, and the Flex sensor. The proposed IoT system is tested in an indoor environment for various scenarios of falls using four models of machine learning. The results from our system showed an accuracy of 0.96%. Compared to the state-of-the-art, our system is simpler and more accurate since it avoids sending false alarms to the blind caretakers.

Optimal 3D Grasp Planning for unknown objects (임의 물체에 대한 최적 3차원 Grasp Planning)

  • 이현기;최상균;이상릉
    • Proceedings of the Korean Society of Precision Engineering Conference
    • /
    • 2002.05a
    • /
    • pp.462-465
    • /
    • 2002
  • This paper deals with the problem of synthesis of stable and optimal grasps with unknown objects by 3-finger hand. Previous robot grasp research has analyzed mainly with either unknown objects 2D by vision sensor or unknown objects, cylindrical or hexahedral objects, 3D. Extending the previous work, in this paper we propose an algorithm to analyze grasp of unknown objects 3D by vision sensor. This is archived by two steps. The first step is to make a 3D geometrical model of unknown objects by stereo matching which is a kind of 3D computer vision technique. The second step is to find the optimal grasping points. In this step, we choose the 3-finger hand because it has the characteristic of multi-finger hand and is easy to modeling. To find the optimal grasping points, genetic algorithm is used and objective function minimizing admissible farce of finger tip applied to the object is formulated. The algorithm is verified by computer simulation by which an optimal grasping points of known objects with different angles are checked.

  • PDF

Implementation of Visual Data Compressor for Vision Sensor of Mobile Robot (이동로봇의 시각센서를 위한 동영상 압축기 구현)

  • Kim Hyung O;Cho Kyoung Su;Baek Moon Yeal;Kee Chang Doo
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.22 no.9 s.174
    • /
    • pp.99-106
    • /
    • 2005
  • In recent years, vision sensors are widely used to mobile robot for navigation or exploration. The analog signal transmission of visual data being used in this area, however, has some disadvantages including noise weakness in view of the data storage. A large amount of data also makes it difficult to use this method for a mobile robot. In this paper, a digital data compressing technology based on MPEG4 which substitutes for analog technology is proposed to overcome the disadvantages by using DWT(Discreate Wavelet Transform) instead of DCT(Discreate Cosine Transform). The TI Company's DSP chip, TMS320C6711, is used for the image encoder, and the performance of the proposed method is evaluated by PSNR(Peake Signal to Noise Rates), QP(Quantization Parameter) and bitrate.

The Optimal Grasp Planning by Using a 3-D Computer Vision Technique (3차원 영상처리 기술을 이용한 Grasp planning의 최적화)

  • 이현기;김성환;최상균;이상룡
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.19 no.11
    • /
    • pp.54-64
    • /
    • 2002
  • This paper deals with the problem of synthesis of stable and optimal grasps with unknown objects by 3-finger hand. Previous robot grasp research has mainly analyzed with either unknown objects 2-dimensionally by vision sensor or known objects, such as cylindrical objects, 3-dimensionally. As extending the previous work, in this study we propose an algorithm to analyze grasp of unknown objects 3-dimensionally by using vision sensor. This is archived by two steps. The first step is to make a 3-dimensional geometrical model for unknown objects by using stereo matching. The second step is to find the optimal grasping points. In this step, we choose the 3-finger hand which has the characteristic of multi-finger hand and is easy to model. To find the optimal grasping points, genetic algorithm is employed and objective function minimizes the admissible force of finger tip applied to the objects. The algorithm is verified by computer simulation by which optimal grasping points of known objects with different angle are checked.

Research about Tool Wear Monitoring in CNC Lathe Machining (선삭 공정에서 공구모니터링에 관한 연구 (I)-공구마모)

  • Go, Jeong-Han;Kim, Yeong-Tae;Lee, Sang-Jo
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.17 no.12
    • /
    • pp.54-60
    • /
    • 2000
  • Research about tool condition monitoring has been done until now for product automation and unmaned system. But it is hard to apply it to the industrial field due to its cost and reliability. This paper presents the new method of tool wear measurement using Marpos gauge. This is a kind of touch sensor, so its cost is lower than vision system. And it is not affected by dust and illumination, which are important in vision system. This proposed method use tool clearance angle to measure flank wear. Experimental results compared with vision system shows that this method is available for tool condition monitoring system.

  • PDF

A Study on Automatic Seam Tracking using Vision Sensor (비전센서를 이용한 자동추적장치에 관한 연구)

  • 전진환;조택동;양상민
    • Proceedings of the Korean Society of Precision Engineering Conference
    • /
    • 1995.10a
    • /
    • pp.1105-1109
    • /
    • 1995
  • A CCD-camera, which is structured with vision system, was used to realize automatic seam-tracking system and 3-D information which is needed to generate torch path, was obtained by using laser-slip beam. To extract laser strip and obtain welding-specific point, Adaptive Hough-transformation was used. Although the basic Hough transformation takes too much time to process image on line, it has a tendency to be robust to the noises as like spatter. For that reson, it was complemented with Adaptive Hough transformation to have an on-line processing ability for scanning a welding-specific point. the dead zone,where the sensing of weld line is impossible, is eliminated by rotating the camera with its rotating axis centered at welding torch. The camera angle is controlled so as to get the minimum image data for the sensing of weld line, hence the image processing time is reduced. The fuzzy controller is adapted to control the camera angle.

  • PDF

Development of a vision sensor for measuring the weld groove parameters in arc welding process (자동 아크 용접공정의 용접개선변수 측정을 위한 시각 시스템)

  • 김호학;부광석;조형석
    • Journal of Welding and Joining
    • /
    • v.8 no.2
    • /
    • pp.58-69
    • /
    • 1990
  • In conventional arc welding, position error of the weld torch with respect to the weld seam and variation of groove dimension are induced by inaccurate fitup and fixturing. In this study, a vision system has been developed to recognize and compensate the position error and dimensional inaccuracy. The system uses a structured laser light illuminated on the weld groove and perceived by a C.C.D camera. A new algorithm to detect the edge of the reflected laser light is introduced for real time processing. The developed system was applied to arbitarary weld paths with various types of joint in arc welding process. The experimental results show that the proposed system can detect the weld groove parameters within good accuracy and yield good tracking performance.

  • PDF

2D Map generation Using Omnidirectional Image sensor and Stereo Vision for MobileRobot MAIRO (자율이동로봇MAIRO의 전방향 이미지센서와 스테레오 비전 시스템을 이용한 2차원 지도 생성)

  • Kim, Kyung-Ho;Lee, Hyung-Kyu;Son, Young-Jun;Song, Jae-Keun
    • Proceedings of the KIEE Conference
    • /
    • 2002.11c
    • /
    • pp.495-500
    • /
    • 2002
  • Recently, a service robot industry outstands as an up and coming industry of the next generation. Specially, there are so many research in self-steering movement(SSM). In order to implement SSM, robot must effectively recognize all around, detect objects and make a surrounding map with sensors. So, many robots have a sonar and a infrared sensor, etc. But, in these sensors, We only know informations about between the robot and the object as well as resolution faculty is of inferior quality. In this paper, we will introduce new algorithm that recognizes objects around robot and makes a two dimension surrounding map with a omni-direction vision camera and two stereo vision cameras.

  • PDF

Human Hand Detection Using Color Vision (컬러 시각을 이용한 사람 손의 검출)

  • Kim, Jun-Yup;Do, Yong-Tae
    • Journal of Sensor Science and Technology
    • /
    • v.21 no.1
    • /
    • pp.28-33
    • /
    • 2012
  • The visual sensing of human hands plays an important part in many man-machine interaction/interface systems. Most existing visionbased hand detection techniques depend on the color cues of human skin. The RGB color image from a vision sensor is often transformed to another color space as a preprocessing of hand detection because the color space transformation is assumed to increase the detection accuracy. However, the actual effect of color space transformation has not been well investigated in literature. This paper discusses a comparative evaluation of the pixel classification performance of hand skin detection in four widely used color spaces; RGB, YIQ, HSV, and normalized rgb. The experimental results indicate that using the normalized red-green color values is the most reliable under different backgrounds, lighting conditions, individuals, and hand postures. The nonlinear classification of pixel colors by the use of a multilayer neural network is also proposed to improve the detection accuracy.