• Title/Summary/Keyword: 템플릿매칭

Search Result 174, Processing Time 0.033 seconds

A Study on the Development of Intelligent Behavior of Humanoid Robot (휴머노이드 로봇의 지능적 행위 구현에 관한 연구)

  • Suh, Joohee;Jang, Inwoo;Woo, Chongwoo
    • Annual Conference of KIPS
    • /
    • 2008.11a
    • /
    • pp.23-26
    • /
    • 2008
  • 본 논문에서는 로봇의 지능적 행위를 구현하기 위하여 인공지능의 몇 가지 기법을 휴머노이드 로봇에 적용하고 이를 테스트 도메인에서 실험하는 연구결과를 기술하였다. 본 연구에서 적용한 기법들은, 인공지능의 계획기법에 기반한 로봇의 계획생성, A* 알고리즘을 적용한 길 찾기, 외부 센서 값에 기반한 장애물회피 및 로봇의 자기 위치인식, 그리고 원하는 물체를 파악하기 위해 템플릿 매칭을 이용한 영상인식 등 네 가지 방향으로 접근하였다. 전반적으로 로봇의 실험은, 웹 페이지로부터 사용자의 쇼핑 목록을 입력 받아, 인공지능의 계획기법에 기반하여 서버에서 이에 대한 실행계획을 만들고 난 후, 로봇이 서버로부터 TCP/IP 기반의 소켓 통신을 통하여 세부 실행계획을 전달받아 임무를 수행하게 된다. 또한 이러한 임무를 수행하기 위해서는 로봇자신의 현재위치에 대한 정보 및 목표물에 대한 위치인식이 요구되며, 이를 위해서 사전에 주어진 맵의 좌표를 찾아가는 방법을 사용하였다.

Iterative Generalized Hough Transform using Multiresolution Search (다중해상도 탐색을 이용한 반복 일반화 허프 변환)

  • ;W. Nick Street
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.10
    • /
    • pp.973-982
    • /
    • 2003
  • This paper presents an efficient method for automatically detecting objects in a given image. The GHT is a robust template matching algorithm for automatic object detection in order to find objects of various shapes. Many different templates are applied by the GHT in order to find objects of various shapes and size. Every boundary detected by the GHT scan be used as an initial outline for more precise contour-finding techniques. The main weakness of the GHT is the excessive time and memory requirements. In order to overcome this drawback, the proposed algorithm uses a multiresolution search by scaling down the original image to half-sized and quarter-sized images. Using the information from the first iterative GHT on a quarter-sized image, the range of nuclear sizes is determined to limit the parameter space of the half-sized image. After the second iterative GHT on the half-sized image, nuclei are detected by the fine search and segmented with edge information which helps determine the exact boundary. The experimental results show that this method gives reduction in computation time and memory usage without loss of accuracy.

Sensor Fusion Docking System of Drone and Ground Vehicles Using Image Object Detection (영상 객체 검출을 이용한 드론과 지상로봇의 센서 융합 도킹 시스템)

  • Beck, Jong-Hwan;Park, Hee-Su;Oh, Se-Ryeong;Shin, Ji-Hun;Kim, Sang-Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.6 no.4
    • /
    • pp.217-222
    • /
    • 2017
  • Recent studies for working robot in dangerous places have been carried out on large unmanned ground vehicles or 4-legged robots with the advantage of long working time, but it is difficult to apply in practical dangerous fields which require the real-time system with high locomotion and capability of delicate working. This research shows the collaborated docking system of drone and ground vehicles which combines image processing algorithm and laser sensors for effective detection of docking markers, and is finally capable of moving a long distance and doing very delicate works. We proposed the docking system of drone and ground vehicles with sensor fusion which also suggests two template matching methods appropriate for this application. The system showed 95% docking success rate in 50 docking attempts.

The Implementation of Automatic Compensation Modules for Digital Camera Image by Recognition of the Eye State (눈의 상태 인식을 이용한 디지털 카메라 영상 자동 보정 모듈의 구현)

  • Jeon, Young-Joon;Shin, Hong-Seob;Kim, Jin-Il
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.14 no.3
    • /
    • pp.162-168
    • /
    • 2013
  • This paper examines the implementation of automatic compensation modules for digital camera image when a person is closing his/her eyes. The modules detect the face and eye region and then recognize the eye state. If the image is taken when a person is closing his/her eyes, the function corrects the eye and produces the image by using the most satisfactory image of the eye state among the past frames stored in the buffer. In order to recognize the face and eye precisely, the pre-process of image correction is carried out using SURF algorithm and Homography method. For the detection of face and eye region, Haar-like feature algorithm is used. To decide whether the eye is open or not, similarity comparison method is used along with template matching of the eye region. The modules are tested in various facial environments and confirmed to effectively correct the images containing faces.

Development of Minutiae-level Compensation Algorithms for Interoperable Fingerprint Recognition (이기종 센서의 호환을 위한 지문 특징점 보정 알고리즘 개발)

  • Jang, Ji-Hyeon;Kim, Hak-Il
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.17 no.5
    • /
    • pp.39-53
    • /
    • 2007
  • The purpose of this paper is the development of a compensation algorithm by which the interoperability of fingerprint recognition can be improved among various different fingerprint sensor. In order to compensate for the different characteristics of fingerprint sensor, an initial evaluation of the sensors using both the ink-stamped method and the flat artificial finger pattern method was undertaken. This paper proposes Common resolution method and Relative resolution method for compensating different resolution of fingerprint images captured by disparate sensors. Both methods can be applied to image-level and minutia-level. In order to compensate the direction of minutiae in minutia-level, Unit vector method is proposed. The EER of the proposed method was improved by average 64.8% better than before compensation. This paper will make a significant contribution to interoperability in the system integration using different sensors.

Container BIC-code region extraction and recognition method using multiple thresholding (다중 이진화를 이용한 컨테이너 BIC 부호 영역 추출 및 인식 방법)

  • Song, Jae-wook;Jung, Na-ra;Kang, Hyun-soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.6
    • /
    • pp.1462-1470
    • /
    • 2015
  • The container BIC-code is a transport protocol for convenience in international shipping and combined transport environment. It is an identification code of a marine transport container which displays a wide variety of information including country's code. Recently, transportation through aircrafts and ships continues to rise. Thus fast and accurate processes are required in the ports to manage transportation. Accordingly, in this paper, we propose a BIC-code region extraction and recognition method using multiple thresholds. In the code recognition, applying a fixed threshold is not reasonable due to a variety of illumination conditions caused by change of weather, lightening, camera position, color of the container and so on. Thus, the proposed method selects the best recognition result at the final stage after applying multiple thresholds to recognition. For each threshold, we performs binarization, labeling, BIC-code pattern decision (horizontal or vertical pattern) by morphological close operation, and character separation from the BIC-code. Then, each characters is recognized by template matching. Finally we measure recognition confidence scores for all the thresholds and choose the best one. Experimental results show that the proposed method yields accurate recognition for the container BIC-code with robustness to illumination change.

Face Detection Using Adaboost and Template Matching of Depth Map based Block Rank Patterns (Adaboost와 깊이 맵 기반의 블록 순위 패턴의 템플릿 매칭을 이용한 얼굴검출)

  • Kim, Young-Gon;Park, Rae-Hong;Mun, Seong-Su
    • Journal of Broadcast Engineering
    • /
    • v.17 no.3
    • /
    • pp.437-446
    • /
    • 2012
  • A face detection algorithms using two-dimensional (2-D) intensity or color images have been studied for decades. Recently, with the development of low-cost range sensor, three-dimensional (3-D) information (i.e., depth image that represents the distance between a camera and objects) can be easily used to reliably extract facial features. Most people have a similar pattern of 3-D facial structure. This paper proposes a face detection method using intensity and depth images. At first, adaboost algorithm using intensity image classifies face and nonface candidate regions. Each candidate region is divided into $5{\times}5$ blocks and depth values are averaged in each block. Then, $5{\times}5$ block rank pattern is constructed by sorting block averages of depth values. Finally, candidate regions are classified as face and nonface regions by matching the constructed depth map based block rank patterns and a template pattern that is generated from training data set. For template matching, the $5{\times}5$ template block rank pattern is prior constructed by averaging block ranks using training data set. The proposed algorithm is tested on real images obtained by Kinect range sensor. Experimental results show that the proposed algorithm effectively eliminates most false positives with true positives well preserved.

Automatic Video Editing Technology based on Matching System using Genre Characteristic Patterns (장르 특성 패턴을 활용한 매칭시스템 기반의 자동영상편집 기술)

  • Mun, Hyejun;Lim, Yangmi
    • Journal of Broadcast Engineering
    • /
    • v.25 no.6
    • /
    • pp.861-869
    • /
    • 2020
  • We introduce the application that automatically makes several images stored in user's device into one video by using the different climax patterns appearing for each film genre. For the classification of the genre characteristics of movies, a climax pattern model style was created by analyzing the genre of domestic movie drama, action, horror and foreign movie drama, action, and horror. The climax pattern was characterized by the change in shot size, the length of the shot, and the frequency of insert use in a specific scene part of the movie, and the result was visualized. The model visualized by genre developed as a template using Firebase DB. Images stored in the user's device were selected and matched with the climax pattern model developed as a template for each genre. Although it is a short video, it is a feature of the proposed application that it can create an emotional story video that reflects the characteristics of the genre. Recently, platform operators such as YouTube and Naver are upgrading applications that automatically generate video using a picture or video taken by the user directly with a smartphone. However, applications that have genre characteristics like movies or include video-generation technology to show stories are still insufficient. It is predicted that the proposed automatic video editing has the potential to develop into a video editing application capable of transmitting emotions.

Extraction and Revision of Building Information from Single High Resolution Image and Digital Map (단일 고해상도 위성영상과 수치지도로부터 건물 정보 추출 및 갱신)

  • Byun, Young-Gi;Kim, Hye-Jin;Choi, Jae-Wan;Han, You-Kyung;Kim, Yong-Il
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.26 no.2
    • /
    • pp.149-156
    • /
    • 2008
  • In this paper, we propose a method aiming at updating the building information of the digital maps using single high resolution satellite image and digital map. Firstly we produced a digital orthoimage through the automatic co-registration of QuickBird image and 1:1,000 digital map. Secondly we extracted building height information through the template matching of digital map's building vector data and the image's edges obtained by Canny operator. Finally we refined the shape of some buildings by using the result from template matching as the seed polygon of the greedy snake algorithm. In order to evaluate the proposed method's effectiveness, we estimated accuracy of the extracted building information using LiDAR DSM and 1:1,000 digital map. The evaluation results showed the proposed method has a good potential for extraction and revision of building information.

Development of a Video Caption Recognition System for Sport Event Broadcasting (스포츠 중계를 위한 자막 인식 시스템 개발)

  • Oh, Ju-Hyun
    • 한국HCI학회:학술대회논문집
    • /
    • 2009.02a
    • /
    • pp.94-98
    • /
    • 2009
  • A video caption recognition system has been developed for broadcasting sport events such as major league baseball. The purpose of the system is to translate the information expressed in English units such as miles per hour (MPH) to the international system of units (SI) such as km/h. The system detects the ball speed displayed in the video and recognizes the numerals. The ball speed is then converted to km/h and displayed by the following character generator (CG) system. Although neural-network based methods are widely used for character and numeral recognition, we use template matching to avoid the training process required before the broadcasting. With the proposed template matching method, the operator can cope with the situation when the caption’s appearance changed without any notification. Templates are configured by the operator with a captured screenshot of the first pitch with ball speed. Templates are updated with following correct recognition results. The accuracy of the recognition module is over 97%, which is still not enough for live broadcasting. When the recognition confidence is low, the system asks the operator for the correct recognition result. The operator chooses the right one using hot keys.

  • PDF