• Title/Summary/Keyword: vision-based technology

Search Result 1,063, Processing Time 0.033 seconds

Vision-Based Piano Music Transcription System (비전 기반 피아노 자동 채보 시스템)

  • Park, Sang-Uk;Park, Si-Hyun;Park, Chun-Su
    • Journal of IKEEE
    • /
    • v.23 no.1
    • /
    • pp.249-253
    • /
    • 2019
  • Most of music-transcription systems that have been commercialized operate based on audio information. However, these conventional systems have disadvantages of environmental dependency, equipment dependency, and time latency. This paper studied a vision-based music-transcription system that utilizes video information rather than audio information, which is a traditional method of music-transcription programs. Computer vision technology is widely used as a field for analyzing and applying information from equipment such as cameras. In this paper, we created a program to generate MIDI file which is electronic music notes by using smart-phone cameras to record the play of piano.

A Study on the Development of Multi-User Virtual Reality Moving Platform Based on Hybrid Sensing (하이브리드 센싱 기반 다중참여형 가상현실 이동 플랫폼 개발에 관한 연구)

  • Jang, Yong Hun;Chang, Min Hyuk;Jung, Ha Hyoung
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.3
    • /
    • pp.355-372
    • /
    • 2021
  • Recently, high-performance HMDs (Head-Mounted Display) are becoming wireless due to the growth of virtual reality technology. Accordingly, environmental constraints on the hardware usage are reduced, enabling multiple users to experience virtual reality within a single space simultaneously. Existing multi-user virtual reality platforms use the user's location tracking and motion sensing technology based on vision sensors and active markers. However, there is a decrease in immersion due to the problem of overlapping markers or frequent matching errors due to the reflected light. Goal of this study is to develop a multi-user virtual reality moving platform in a single space that can resolve sensing errors and user immersion decrease. In order to achieve this goal hybrid sensing technology was developed, which is the convergence of vision sensor technology for position tracking, IMU (Inertial Measurement Unit) sensor motion capture technology and gesture recognition technology based on smart gloves. In addition, integrated safety operation system was developed which does not decrease the immersion but ensures the safety of the users and supports multimodal feedback. A 6 m×6 m×2.4 m test bed was configured to verify the effectiveness of the multi-user virtual reality moving platform for four users.

Omni-directional Vision SLAM using a Motion Estimation Method based on Fisheye Image (어안 이미지 기반의 움직임 추정 기법을 이용한 전방향 영상 SLAM)

  • Choi, Yun Won;Choi, Jeong Won;Dai, Yanyan;Lee, Suk Gyu
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.8
    • /
    • pp.868-874
    • /
    • 2014
  • This paper proposes a novel mapping algorithm in Omni-directional Vision SLAM based on an obstacle's feature extraction using Lucas-Kanade Optical Flow motion detection and images obtained through fish-eye lenses mounted on robots. Omni-directional image sensors have distortion problems because they use a fish-eye lens or mirror, but it is possible in real time image processing for mobile robots because it measured all information around the robot at one time. In previous Omni-Directional Vision SLAM research, feature points in corrected fisheye images were used but the proposed algorithm corrected only the feature point of the obstacle. We obtained faster processing than previous systems through this process. The core of the proposed algorithm may be summarized as follows: First, we capture instantaneous $360^{\circ}$ panoramic images around a robot through fish-eye lenses which are mounted in the bottom direction. Second, we remove the feature points of the floor surface using a histogram filter, and label the candidates of the obstacle extracted. Third, we estimate the location of obstacles based on motion vectors using LKOF. Finally, it estimates the robot position using an Extended Kalman Filter based on the obstacle position obtained by LKOF and creates a map. We will confirm the reliability of the mapping algorithm using motion estimation based on fisheye images through the comparison between maps obtained using the proposed algorithm and real maps.

A Real-time Vehicle Localization Algorithm for Autonomous Parking System (자율 주차 시스템을 위한 실시간 차량 추출 알고리즘)

  • Hahn, Jong-Woo;Choi, Young-Kyu
    • Journal of the Semiconductor & Display Technology
    • /
    • v.10 no.2
    • /
    • pp.31-38
    • /
    • 2011
  • This paper introduces a video based traffic monitoring system for detecting vehicles and obstacles on the road. To segment moving objects from image sequence, we adopt the background subtraction algorithm based on the local binary patterns (LBP). Recently, LBP based texture analysis techniques are becoming popular tools for various machine vision applications such as face recognition, object classification and so on. In this paper, we adopt an extension of LBP, called the Diagonal LBP (DLBP), to handle the background subtraction problem arise in vision-based autonomous parking systems. It reduces the code length of LBP by half and improves the computation complexity drastically. An edge based shadow removal and blob merging procedure are also applied to the foreground blobs, and a pose estimation technique is utilized for calculating the position and heading angle of the moving object precisely. Experimental results revealed that our system works well for real-time vehicle localization and tracking applications.

Implementation of Visual Data Compressor for Vision Sensor of Mobile Robot (이동로봇의 시각센서를 위한 동영상 압축기 구현)

  • Kim Hyung O;Cho Kyoung Su;Baek Moon Yeal;Kee Chang Doo
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.22 no.9 s.174
    • /
    • pp.99-106
    • /
    • 2005
  • In recent years, vision sensors are widely used to mobile robot for navigation or exploration. The analog signal transmission of visual data being used in this area, however, has some disadvantages including noise weakness in view of the data storage. A large amount of data also makes it difficult to use this method for a mobile robot. In this paper, a digital data compressing technology based on MPEG4 which substitutes for analog technology is proposed to overcome the disadvantages by using DWT(Discreate Wavelet Transform) instead of DCT(Discreate Cosine Transform). The TI Company's DSP chip, TMS320C6711, is used for the image encoder, and the performance of the proposed method is evaluated by PSNR(Peake Signal to Noise Rates), QP(Quantization Parameter) and bitrate.

Sorting for Plastic Bottles Recycling using Machine Vision Methods

  • SanaSadat Mirahsani;Sasan Ghasemipour;AmirAbbas Motamedi
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.6
    • /
    • pp.89-98
    • /
    • 2024
  • Due to the increase in population and consequently the increase in the production of plastic waste, recovery of this part of the waste is an undeniable necessity. On the other hand, the recycling of plastic waste, if it is placed in a systematic process and controlled, can be effective in creating jobs and maintaining environmental health. Waste collection in many large cities has become a major problem due to lack of proper planning with increasing waste from population accumulation and changing consumption patterns. Today, waste management is no longer limited to waste collection, but waste collection is one of the important areas of its management, i.e. training, segregation, collection, recycling and processing. In this study, a systematic method based on machine vision for sorting plastic bottles in different colors for recycling purposes will be proposed. In this method, image classification and segmentation techniques were presented to improve the performance of plastic bottle classification. Evaluation of the proposed method and comparison with previous works showed the proper performance of this method.

Development Technology of Vision Based Augmented Reality for the Maintenance of Products (제품의 유지보수를 위한 시각 기반 증강현실 기술 개발)

  • Lee, Kyung-Ho;Lee, Jung-Min;Kim, Dong-Geun;Han, Yung-Su;Lee, Jea-Jun
    • Korean Journal of Computational Design and Engineering
    • /
    • v.13 no.4
    • /
    • pp.265-272
    • /
    • 2008
  • The flow of technology is going to human-oriented direction, from the time when the computer was first invented, to now where new computing environment using mobile and global network are everywhere. On that technology flow, ubiquitous is being suggested as new paradigm of computing environment. Augmented Reality is one of ubiquitous technologies that provide the interactions between human and computer. By adding computer-generated information to real information and their interaction, user can get the improved and more knowledgeable information about real world. The purpose of this paper is to show the possibility of applying vision based augmented reality to maintenance of product system.

A Video Traffic Flow Detection System Based on Machine Vision

  • Wang, Xin-Xin;Zhao, Xiao-Ming;Shen, Yu
    • Journal of Information Processing Systems
    • /
    • v.15 no.5
    • /
    • pp.1218-1230
    • /
    • 2019
  • This study proposes a novel video traffic flow detection method based on machine vision technology. The three-frame difference method, which is one kind of a motion evaluation method, is used to establish initial background image, and then a statistical scoring strategy is chosen to update background image in real time. Finally, the background difference method is used for detecting the moving objects. Meanwhile, a simple but effective shadow elimination method is introduced to improve the accuracy of the detection for moving objects. Furthermore, the study also proposes a vehicle matching and tracking strategy by combining characteristics, such as vehicle's location information, color information and fractal dimension information. Experimental results show that this detection method could quickly and effectively detect various traffic flow parameters, laying a solid foundation for enhancing the degree of automation for traffic management.

Real-Time Fire Detection Method Using YOLOv8 (YOLOv8을 이용한 실시간 화재 검출 방법)

  • Tae Hee Lee;Chun-Su Park
    • Journal of the Semiconductor & Display Technology
    • /
    • v.22 no.2
    • /
    • pp.77-80
    • /
    • 2023
  • Since fires in uncontrolled environments pose serious risks to society and individuals, many researchers have been investigating technologies for early detection of fires that occur in everyday life. Recently, with the development of deep learning vision technology, research on fire detection models using neural network backbones such as Transformer and Convolution Natural Network has been actively conducted. Vision-based fire detection systems can solve many problems with physical sensor-based fire detection systems. This paper proposes a fire detection method using the latest YOLOv8, which improves the existing fire detection method. The proposed method develops a system that detects sparks and smoke from input images by training the Yolov8 model using a universal fire detection dataset. We also demonstrate the superiority of the proposed method through experiments by comparing it with existing methods.

  • PDF

Robust Defect Size Measuring Method for an Automated Vision Inspection System (영상기반 자동결함 검사시스템에서 재현성 향상을 위한 결함 모델링 및 측정 기법)

  • Joo, Young-Bok;Huh, Kyung-Moo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.11
    • /
    • pp.974-978
    • /
    • 2013
  • AVI (Automatic Vision Inspection) systems automatically detect defect features and measure their sizes via camera vision. AVI systems usually report different measurements on the same defect with some variations on position or rotation mainly because different images are provided. This is caused by possible variations from the image acquisition process including optical factors, nonuniform illumination, random noises, and so on. For this reason, conventional area based defect measuring methods have problems of robustness and consistency. In this paper, we propose a new defect size measuring method to overcome this problem, utilizing volume information that is completely ignored in the area based defect measuring method. The results show that our proposed method dramatically improves the robustness and consistency of defect size measurement.