• 제목/요약/키워드: Character segmentation

검색결과 172건 처리시간 0.03초

한글 인쇄체 문자인식 전용 신경망 Coprocessor의 구현에 관한 연구 (Study on Implementation of a neural Coprocessor for Printed Hangul-Character Recognition)

  • 김영철;이태원
    • 한국정보처리학회논문지
    • /
    • 제5권1호
    • /
    • pp.119-127
    • /
    • 1998
  • 본 논문에서는 한글 인쇄체 인식 시스템의 실시간 처리를 위하여 인식 프로세스중 시간이 많이 걸리는 한글 문자 유형 분류 및 자소 인식 단계를 고속 처리할 수 있는 다층구조 신경망을 VLSI 설계 하였으며, 신경망과 호스트 컴퓨터간의 인터페이스와 신경망 제어를 담당하는 코프로세서 구조를 제안하였다. 이를 VHDL 모델링 및 논리합성을 통하여 설계하여 시뮬레이션을 통하여 구조와 동작 및 성능을 검증하였다. 실험결과 제안한 신경망 coprocessor는 기존의 소프트웨어 구현 인식 시스템의 유형 분류 및 자소 인식률과 대등한 성능을 보인 반면 고속의 인식속도를 보였다.

  • PDF

방향 정규화 및 CNN 딥러닝 기반 차량 번호판 인식에 관한 연구 (A Study on the License Plate Recognition Based on Direction Normalization and CNN Deep Learning)

  • 기재원;조성원
    • 한국멀티미디어학회논문지
    • /
    • 제25권4호
    • /
    • pp.568-574
    • /
    • 2022
  • In this paper, direction normalization and CNN deep learning are used to develop a more reliable license plate recognition system. The existing license plate recognition system consists of three main modules: license plate detection module, character segmentation module, and character recognition module. The proposed system minimizes recognition error by adding a direction normalization module when a detected license plate is inclined. Experimental results show the superiority of the proposed method in comparison to the previous system.

A Fast Algorithm for Korean Text Extraction and Segmentation from Subway Signboard Images Utilizing Smartphone Sensors

  • Milevskiy, Igor;Ha, Jin-Young
    • Journal of Computing Science and Engineering
    • /
    • 제5권3호
    • /
    • pp.161-166
    • /
    • 2011
  • We present a fast algorithm for Korean text extraction and segmentation from subway signboards using smart phone sensors in order to minimize computational time and memory usage. The algorithm can be used as preprocessing steps for optical character recognition (OCR): binarization, text location, and segmentation. An image of a signboard captured by smart phone camera while holding smart phone by an arbitrary angle is rotated by the detected angle, as if the image was taken by holding a smart phone horizontally. Binarization is only performed once on the subset of connected components instead of the whole image area, resulting in a large reduction in computational time. Text location is guided by user's marker-line placed over the region of interest in binarized image via smart phone touch screen. Then, text segmentation utilizes the data of connected components received in the binarization step, and cuts the string into individual images for designated characters. The resulting data could be used as OCR input, hence solving the most difficult part of OCR on text area included in natural scene images. The experimental results showed that the binarization algorithm of our method is 3.5 and 3.7 times faster than Niblack and Sauvola adaptive-thresholding algorithms, respectively. In addition, our method achieved better quality than other methods.

접촉점에서의 국소 그래프 패턴에 의한 필기체 한글의 자소분리에 관한 연구 (A Study on the Phoneme Segmentation of Handwritten Korean Characters by Local Graph Patterns on Contacting Points)

  • 최필웅;이기영;구하성;고형화
    • 전자공학회논문지B
    • /
    • 제30B권4호
    • /
    • pp.1-10
    • /
    • 1993
  • In this paper, a new method of phoneme segmentation of handwritten Korean characters using the local graph pattern is proposed. At first, thinning was performed before extracting features. End-point, inflexion-point, branch-point and cross-point were extracted as features. Using these features and the angular relations between these features, local graph pattern was made. When local graph pattern is made, the of strokes is investigated on contacting point. From this process, pattern is simplified as contacting pattern of the basic form and the contacting form we must take into account can be restricted within fixed region, 4therefore phoneme segmentation not influenced by characters form and any other contact in a single character is performed as matching this local graph pattern with base patterns searched ahead. This experiments with 540 characters have been conducted. From the result of this experiment, it is shown that phoneme segmentation is independent of characters form and other contact in a single character to obtain a correct segmentation rate of 95%, manages it efficiently to reduce the time spent in lock operation when the lock.

  • PDF

적응 휴리스틱 분할 알고리즘을 이용한 실시간 차량 번호판 인식 시스템 (Real-Time Vehicle License Plate Recognition System Using Adaptive Heuristic Segmentation Algorithm)

  • 진문용;박종빈;이동석;박동선
    • 정보처리학회논문지:소프트웨어 및 데이터공학
    • /
    • 제3권9호
    • /
    • pp.361-368
    • /
    • 2014
  • 차량 번호판 인식 시스템은 복잡한 교통환경의 효율적 관리를 위해 발전되어 현재 많은 곳에 사용되고 있다. 그러나 조명, 잡음, 배경변화, 번호판 훼손 등 환경변화에 큰 영향을 받기 때문에 제한된 환경에서만 동작하며, 실시간으로 사용하기 어렵다. 본 논문에서는 조명변화와 잡음에 강건하며 빠른 번호판 인식을 위한 휴리스틱 분할 알고리즘 및 이를 이용한 실시간 번호판 인식 시스템을 제안한다. 첫 번째 단계는 Haar-like 특징과 Adaboost를 이용하여 번호판을 검출한다. 이 방법은 적분영상을 이용하며 케스케이드 구조로 구성되어 있어 빠른 검출이 가능하다. 두 번째 단계에서 적응 히스토그램 평활화 방법과 노이즈를 경감시키는 바이레터럴 필터를 이용하여 번호판의 종류를 결정한 후, 번호판 종류에 따라 적분영상을 이용한 적응 이진화, 픽셀 프로젝션, 사전지식 등을 기반으로 빠르고 정확한 문자 분할을 한다. 세번째 단계에서는 HOG와 신경망 알고리즘을 이용하여 숫자를 인식하고, SVM을 이용해 한글을 인식한다. 실험결과는 번호판검출에 94.29%의 검출률, 2.94%의 오경보율을 보이며, 문자분할에서는 검출률 97.23%, 2.94%의 오경보율을 보였다. 문자인식에서 평균 인식률은 98.38%이다. 평균 운용시간은 140ms으로 빠르고 강인한 실시간 시스템을 만들 수 있다.

자동차 VIN 문자 인식 시스템 개발 (Development of VIN Character Recognition System for Motor)

  • 이용중;이화춘;류재엽
    • 한국공작기계학회:학술대회논문집
    • /
    • 한국공작기계학회 2000년도 추계학술대회논문집 - 한국공작기계학회
    • /
    • pp.68-73
    • /
    • 2000
  • This study to embody automatic recognition of VIN(Vehicle Identification Number)character by computer vision system. Automatic recognition characters methods consist of the thining processing and the recognition of each character. VIN character and background classified using counting method of the size of connected pixels. Thining processing applied to segmentation of connected fundamental phonemes by Hilditch's algorithm. Each VIN character contours tracing algorithm used the Freeman's direction tracing algorithm.

  • PDF

칼라 문서에서 문자 영역 추출믹 문자분리 (The Character Area Extraction and the Character Segmentation on the Color Document)

  • 김의정
    • 한국지능시스템학회논문지
    • /
    • 제9권4호
    • /
    • pp.444-450
    • /
    • 1999
  • 본 논문에서는 칼라로 입력된 문서 영상에서 문자 영역추출을 위하여 k-means을 이용한 클러스트링 알고리즘을 제안하였다. 칼라 영상의 클러스트링을 위해서 HIS 좌표계에 적합한 거리함수를 제안하였다. 이를 인식하기 위한 전처리 단계인 문자분리(segmentation)방법은 연결 화소를 이용한 개별문자 추출 알고리즘을 제안하였다. 본 알고리즘 에서는 문자분리방벙에서 접촉문자 (touching character)또는 겹친 문자(overlapped character)등과 같이 분리가 곤란한 문자를 개별문자로 분리하는 방법이다. 기존의 문자 분리방법에서는 투영(projection)dop 의한 방법과 외곽선(edge)추적에 의한 방법등을 사용하여 왔으나 제안된 방법은 문자열 추출후 한번의 투영으로 연결화소를 이용하여 개별문자를 추출한다. 문자 영역과 비 문자 영역을 구분하여 개발문자 추출을 한 결과 단순한 이진 영상이 아닌 칼라 영상에서의 문서 처리가 큰 의의가 있고 기존의 문서 처리기 보다 향상된 알고리즘인 것을 확인하였다.

  • PDF

문서영상의 에지 정보를 이용한 효과적인 블록분할 및 유형분류 (An Efficient Block Segmentation and Classification of a Document Image Using Edge Information)

  • 박창준;전준형;최형문
    • 전자공학회논문지B
    • /
    • 제33B권10호
    • /
    • pp.120-129
    • /
    • 1996
  • This paper presents an efficient block segmentation and classification using the edge information of the document image. We extract four prominent features form the edge gradient and orientaton, all of which, and thereby the block clssifications, are insensitive to the background noise and the brightness variation of of the image. Using these four features, we can efficiently classify a document image into the seven categrories of blocks of small-size letters, large-size letters, tables, equations, flow-charts, graphs, and photographs, the first five of which are text blocks which are character-recognizable, and the last two are non-character blocks. By introducing the clumn interval and text line intervals of the document in the determination of th erun length of CRLA (constrained run length algorithm), we can obtain an efficient block segmentation with reduced memory size. The simulation results show that the proposed algorithm can rigidly segment and classify the blocks of the documents into the above mentioned seven categories and classification performance is high enough for all the categories except for the graphs with too much variations.

  • PDF

Segmentation Algorithm for Wafer ID using Active Multiple Templates Model

  • Ahn, In-Mo;Kang, Dong-Joong;Chung, Yoon-Tack
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2003년도 ICCAS
    • /
    • pp.839-844
    • /
    • 2003
  • This paper presents a method to segment wafer ID marks on poor quality images under uncontrolled lighting conditions of the semiconductor process. The active multiple templates matching method is suggested to search ID areas on wafers and segment them into meaningful regions and it would have been impossible to recognize characters using general OCR algorithms. This active template model is designed by applying a snake model that is used for active contour tracking. Active multiple template model searches character areas and segments them into single characters optimally, tracking each character that can vary in a flexible manner according to string configurations. Applying active multiple templates, the optimization of the snake energy is done using Greedy algorithm, to maximize its efficiency by automatically controlling each template gap. These vary according to the configuration of character string. Experimental results using wafer images from real FA environment are presented.

  • PDF

A Method for Caption Segmentation using Minimum Spanning Tree

  • Chun, Byung-Tae;Kim, Kyuheon;Lee, Jae-Yeon
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2000년도 ITC-CSCC -2
    • /
    • pp.906-909
    • /
    • 2000
  • Conventional caption extraction methods use the difference between frames or color segmentation methods from the whole image. Because these methods depend heavily on heuristics, we should have a priori knowledge of the captions to be extracted. Also they are difficult to implement. In this paper, we propose a method that uses little heuristics and simplified algorithm. We use topographical features of characters to extract the character points and use KMST(Kruskal minimum spanning tree) to extract the candidate regions for captions. Character regions are determined by testing several conditions and verifying those candidate regions. Experimental results show that the candidate region extraction rate is 100%, and the character region extraction rate is 98.2%. And then we can see the results that caption area in complex images is well extracted.

  • PDF