• Title/Summary/Keyword: Camera-based Recognition

Search Result 593, Processing Time 0.034 seconds

Text Region Extraction and OCR on Camera Based Images (카메라 영상 위에서의 문자 영역 추출 및 OCR)

  • Shin, Hyun-Kyung
    • The KIPS Transactions:PartD
    • /
    • v.17D no.1
    • /
    • pp.59-66
    • /
    • 2010
  • Traditional OCR engines are designed to the scanned documents in calibrated environment. Three dimensional perspective distortion and smooth distortion in images are critical problems caused by un-calibrated devices, e.g. image from smart phones. To meet the growing demand of character recognition of texts embedded in the photos acquired from the non-calibrated hand-held devices, we address the problem in three categorical aspects: rotational invariant method of text region extraction, scale invariant method of text line segmentation, and three dimensional perspective mapping. With the integration of the methods, we developed an OCR for camera-captured images.

A Study On Low-cost LPR(License Plate Recognition) System Based On Smart Cam System using Android (안드로이드 기반 스마트 캠 방식의 저가형 자동차 번호판 인식 시스템 구현에 관한 연구)

  • Lee, Hee-Yeol;Lee, Seung-Ho
    • Journal of IKEEE
    • /
    • v.18 no.4
    • /
    • pp.471-477
    • /
    • 2014
  • In this paper, we propose a low-cost license plate recognition system based on smart cam system using Android. The proposed system consists of a portable device and server. Potable device Hardware consists of ARM Cortex-A9 (S5PV210) processor control unit, a power supply device, wired and wireless communication, input/output unit. We develope Linux kernel and dedicated device driver for WiFi module and camera. The license plate recognition algorithm is consisted of setting candidate plates areas with canny edge detector, extracting license plate number with Labeling, recognizing with template matching, etc. The number that is recognized by the device is transmitted to the remote server via the user mobile phone, and the server re-transfer the vehicle information in the database to the portable device. To verify the utility of the proposed system, user photographs the license plate of any vehicle in the natural environment. Confirming the recognition result, the recognition rate was 95%. The proposed system was suitable for low cost portable license plate recognition device, it enabled the stability of the system when used long time by using the Android operating system.

Development of a Bicycle Fitting System Based on Depth Camera through Body Part Recognition (인체부위 인식을 통한 깊이 카메라 기반의 자전거 피팅 시스템 개발)

  • Jeon, Hyesung;Lee, Jinwon;Yang, Jeongsam
    • Korean Journal of Computational Design and Engineering
    • /
    • v.20 no.4
    • /
    • pp.375-384
    • /
    • 2015
  • Recently, there has been a gradual increase in the number of people who are interested in cycling, leading to an increasing number of cycling equipment consumers. However, many bicyclists get hurt because of their lack of knowledge about the right size of bicycle for their body. Although it is necessary for a rider to fit their bicycle to prevent injury, they reject a fitting service because of the long hours and high cost. In this study, we propose a bicycle fitting system that uses a depth camera to improve the limitations of existing manual fitting systems. With the defined formula, the system calculates the size of the bicycle using body image information extracted by a depth camera and visualizes a customized bicycle for a specific consumer. This system will not only save the customer time and money, but will prevent injury from the use of a bicycle that does not fit.

Efficient Object Tracking System Using the Fusion of a CCD Camera and an Infrared Camera (CCD카메라와 적외선 카메라의 융합을 통한 효과적인 객체 추적 시스템)

  • Kim, Seung-Hun;Jung, Il-Kyun;Park, Chang-Woo;Hwang, Jung-Hoon
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.17 no.3
    • /
    • pp.229-235
    • /
    • 2011
  • To make a robust object tracking and identifying system for an intelligent robot and/or home system, heterogeneous sensor fusion between visible ray system and infrared ray system is proposed. The proposed system separates the object by combining the ROI (Region of Interest) estimated from two different images based on a heterogeneous sensor that consolidates the ordinary CCD camera and the IR (Infrared) camera. Human's body and face are detected in both images by using different algorithms, such as histogram, optical-flow, skin-color model and Haar model. Also the pose of human body is estimated from the result of body detection in IR image by using PCA algorithm along with AdaBoost algorithm. Then, the results from each detection algorithm are fused to extract the best detection result. To verify the heterogeneous sensor fusion system, few experiments were done in various environments. From the experimental results, the system seems to have good tracking and identification performance regardless of the environmental changes. The application area of the proposed system is not limited to robot or home system but the surveillance system and military system.

Implementation of Enhanced Vision for an Autonomous Map-based Robot Navigation

  • Roland, Cubahiro;Choi, Donggyu;Kim, Minyoung;Jang, Jongwook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.41-43
    • /
    • 2021
  • Robot Operating System (ROS) has been a prominent and successful framework used in robotics business and academia.. However, the framework has long been focused and limited to navigation of robots and manipulation of objects in the environment. This focus leaves out other important field such as speech recognition, vision abilities, etc. Our goal is to take advantage of ROS capacity to integrate additional libraries of programming functions aimed at real-time computer vision with a depth-image camera. In this paper we will focus on the implementation of an upgraded vision with the help of a depth camera which provides a high quality data for a much enhanced and accurate understanding of the environment. The varied data from the cameras are then incorporated in ROS communication structure for any potential use. For this particular case, the system will use OpenCV libraries to manipulate the data from the camera and provide a face-detection capabilities to the robot, while navigating an indoor environment. The whole system has been implemented and tested on the latest technologies of Turtlebot3 and Raspberry Pi4.

  • PDF

Performance Analysis of Face Recognition by Face Image resolutions using CNN without Backpropergation and LDA (역전파가 제거된 CNN과 LDA를 이용한 얼굴 영상 해상도별 얼굴 인식률 분석)

  • Moon, Hae-Min;Park, Jin-Won;Pan, Sung Bum
    • Smart Media Journal
    • /
    • v.5 no.1
    • /
    • pp.24-29
    • /
    • 2016
  • To satisfy the needs of high-level intelligent surveillance system, it shall be able to extract objects and classify to identify precise information on the object. The representative method to identify one's identity is face recognition that is caused a change in the recognition rate according to environmental factors such as illumination, background and angle of camera. In this paper, we analyze the robust face recognition of face image by changing the distance through a variety of experiments. The experiment was conducted by real face images of 1m to 5m. The method of face recognition based on Linear Discriminant Analysis show the best performance in average 75.4% when a large number of face images per one person is used for training. However, face recognition based on Convolution Neural Network show the best performance in average 69.8% when the number of face images per one person is less than five. In addition, rate of low resolution face recognition decrease rapidly when the size of the face image is smaller than $15{\times}15$.

A method for image-based shadow interaction with virtual objects

  • Ha, Hyunwoo;Ko, Kwanghee
    • Journal of Computational Design and Engineering
    • /
    • v.2 no.1
    • /
    • pp.26-37
    • /
    • 2015
  • A lot of researchers have been investigating interactive portable projection systems such as a mini-projector. In addition, in exhibition halls and museums, there is a trend toward using interactive projection systems to make viewing more exciting and impressive. They can also be applied in the field of art, for example, in creating shadow plays. The key idea of the interactive portable projection systems is to recognize the user's gesture in real-time. In this paper, a vision-based shadow gesture recognition method is proposed for interactive projection systems. The gesture recognition method is based on the screen image obtained by a single web camera. The method separates only the shadow area by combining the binary image with an input image using a learning algorithm that isolates the background from the input image. The region of interest is recognized with labeling the shadow of separated regions, and then hand shadows are isolated using the defect, convex hull, and moment of each region. To distinguish hand gestures, Hu's invariant moment method is used. An optical flow algorithm is used for tracking the fingertip. Using this method, a few interactive applications are developed, which are presented in this paper.

Robust Recognition of 3D Object Using Attributed Relation Graph of Silhouette's (실루엣 기반의 관계그래프 이용한 강인한 3차원 물체 인식)

  • Kim, Dae-Woong;Baek, Kyung-Hwan;Hahn, Hern-Soo
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.25 no.7
    • /
    • pp.103-110
    • /
    • 2008
  • This paper presents a new approach of recognizing a 3D object using a single camera, based on the extended convex hull of its silhouette. It aims at minimizing the DB size and simplifying the processes for matching and feature extraction. For this purpose, two concepts are introduced: extended convex hull and measurable region. Extended convex hull consists of convex curved edges as well as convex polygons. Measurable region is the cluster of the viewing vectors of a camera represented as the points on the orientation sphere from which a specific set of surfaces can be measured. A measurable region is represented by the extended convex hull of the silhouette which can be obtained by viewing the object from the center of the measurable region. Each silhouette is represented by a relation graph where a node describes an edge using its type, length, reality, and components. Experimental results are included to show that the proposed algorithm works efficiently even when the objects are overlapped and partially occluded. The time complexity for searching the object model in the database is O(N) where N is the number of silhouette models.

Development of the UPC-A Barcode Recognition Algorithm for Smartphone Applications (스마트 폰 어플리케이션 적용을 위한 UPC-A Bar code 인식 알고리즘 개발)

  • Lee, Sang-Joon;Lee, Sang-Yong;Lee, Young-Bum;Lee, Myoung-Ho
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.60 no.1
    • /
    • pp.174-183
    • /
    • 2011
  • This paper is about a bar code decoding algorithm developed for smart phone applications. The algorithm consists of bar code data extraction procedure, bar code signal estimation procedure, and bar code decoding procedure. To detect the peak bar code module, a DSTW had been applied because of its outstanding performance in ECG peak detection. In order to minimize errors due to non-uniform light effect, the proposed algorithm was acted as a baseline wandering filter based on module peaks detection. The algorithm had been tested to evaluate the performance under the conditions of blurring, non-uniformed light and white noises. The algorithm had shown excellent performance in reconstruction of bar code decoding, compared to other conventional methods. In order to show the possibility of applying the algorithm to a smart phone, a real UPC-A type 150 bar code pictures obtained from a smart phone camera was applied to the algorithm, achieving the correct recognition rate of 97.33%.

Quality Enhancement of 3D Volumetric Contents Based on 6DoF for 5G Telepresence Service

  • Byung-Seo Park;Woosuk Kim;Jin-Kyum Kim;Dong-Wook Kim;Young-Ho Seo
    • Journal of Web Engineering
    • /
    • v.21 no.3
    • /
    • pp.729-750
    • /
    • 2022
  • In general, the importance of 6DoF (degree of freedom) 3D (dimension) volumetric contents technology is emerging in 5G (generation) telepresence service, Web-based (WebGL) graphics, computer vision, robotics, and next-generation augmented reality. Since it is possible to acquire RGB images and depth images in real-time through depth sensors that use various depth acquisition methods such as time of flight (ToF) and lidar, many changes have been made in object detection, tracking, and recognition research. In this paper, we propose a method to improve the quality of 3D models for 5G telepresence by processing images acquired through depth and RGB cameras on a multi-view camera system. In this paper, the quality is improved in two major ways. The first concerns the shape of the 3D model. A method of removing noise outside the object by applying a mask obtained from a color image and a combined filtering operation to obtain the difference in depth information between pixels inside the object were proposed. Second, we propose an illumination compensation method for images acquired through a multi-view camera system for photo-realistic 3D model generation. It is assumed that the three-dimensional volumetric shooting is done indoors, and the location and intensity of illumination according to time are constant. Since the multi-view camera uses a total of 8 pairs and converges toward the center of space, the intensity and angle of light incident on each camera are different even if the illumination is constant. Therefore, all cameras take a color correction chart and use a color optimization function to obtain a color conversion matrix that defines the relationship between the eight acquired images. Using this, the image input from all cameras is corrected based on the color correction chart. It was confirmed that the quality of the 3D model could be improved by effectively removing noise due to the proposed method when acquiring images of a 3D volumetric object using eight cameras. It has been experimentally proven that the color difference between images is reduced.