• Title/Summary/Keyword: recognition rates

Search Result 524, Processing Time 0.02 seconds

The Design and Implementation of a Performance Evaluation Tool for the Face Recognition System (얼굴인식시스템 성능평가 도구의 설계 및 구현)

  • Shin, Woo-Chang
    • Journal of Information Technology Services
    • /
    • v.6 no.2
    • /
    • pp.161-175
    • /
    • 2007
  • Face recognition technology has lately attracted considerable attention because of its non-intrusiveness, usability and applicability. Related companies insist that their commercial products show the recognition rates more than 95% according to their self-testing. But, the rates cannot be admitted as official recognition rates. So, performance evaluation methods and tools are necessary to objectively measure the accuracy and performance of face recognition systems. In this paper, I propose a reference model for biometrics recognition evaluation tools, and implement an evaluation tool for the face recognition system based on the proposed reference model.

The Recognition of Unvoiced Consonants Using Characteristic Parameters of the Phonemes (음소 특정 파라미터를 이용한 무성자음 인식)

  • 허만택;이종혁;남기곤;윤태훈;김재창;이양성
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.4
    • /
    • pp.175-182
    • /
    • 1994
  • In this study, we present unvoiced consonant recognition system using characteristic parameters of the phoneme of the each syllable. For the recognition, the characteristic parameters on the time domain such as ZCR, total energy of the consonant region and half region energy of the consonant region, and those on the frequency domain such as the frequency spectrum of the transition region are used. The objective unvoiced consonants in this study are /ㄱ/,/ㄷ/,/ㅂ/,/ㅈ/,/ㅋ/,/ㅌ/,/ㅍ/ and /ㅊ/. Each characteristic parameter of two regions extracted from these segmented unvoiced consonants are used for each recognition system of the region, independently, And complementing two outputs of each other system, the final output is to be produced. The recognition system is implemented using MLP which has learning ability. The recognition simulation results for 112 unvoiced consonant samples are that average recognition rates are 96.4$\%$ under 80$\%$ learning rates and 93.7$\%$ under 60$\%$ learning rates.

  • PDF

Presentation Attacks in Palmprint Recognition Systems

  • Sun, Yue;Wang, Changkun
    • Journal of Multimedia Information System
    • /
    • v.9 no.2
    • /
    • pp.103-112
    • /
    • 2022
  • Background: A presentation attack places the printed image or displayed video at the front of the sensor to deceive the biometric recognition system. Usually, presentation attackers steal a genuine user's biometric image and use it for presentation attack. In recent years, reconstruction attack and adversarial attack can generate high-quality fake images, and have high attack success rates. However, their attack rates degrade remarkably after image shooting. Methods: In order to comprehensively analyze the threat of presentation attack to palmprint recognition system, this paper makes six palmprint presentation attack datasets. The datasets were tested on texture coding-based recognition methods and deep learning-based recognition methods. Results and conclusion: The experimental results show that the presentation attack caused by the leakage of the original image has a high success rate and a great threat; while the success rates of reconstruction attack and adversarial attack decrease significantly.

Feature Extraction and Statistical Pattern Recognition for Image Data using Wavelet Decomposition

  • Kim, Min-Soo;Baek, Jang-Sun
    • Communications for Statistical Applications and Methods
    • /
    • v.6 no.3
    • /
    • pp.831-842
    • /
    • 1999
  • We propose a wavelet decomposition feature extraction method for the hand-written character recognition. Comparing the recognition rates of which methods with original image features and with selected features by the wavelet decomposition we study the characteristics of the proposed method. LDA(Linear Discriminant Analysis) QDA(Quadratic Discriminant Analysis) RDA(Regularized Discriminant Analysis) and NN(Neural network) are used for the calculation of recognition rates. 6000 hand-written numerals from CENPARMI at Concordia University are used for the experiment. We found that the set of significantly selected wavelet decomposed features generates higher recognition rate than the original image features.

  • PDF

A Frequency Weighted HMM with Spectral Compensation for Noisy Speech Recognition (잡음하의 음성인식을 위한 스펙트럴 보상과 주파수 가중 HMM)

  • 이광석
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.5 no.3
    • /
    • pp.443-449
    • /
    • 2001
  • This paper is simulation research to improve speech recognition rates under the noisy environment. We examines recognition ratio based on frequency-weighted HMM together with spectral subtraction. As results, frequency-weighted HMM with scaling coefficients is trained as a minimum error classification criterion, and is presents a higher recognition rates in noisy condition than a conventional method. Furthermore, spectral subtraction method gives 11 to 28% improvements for this frequency-weighted HMM in low SNR, and gives recognition rates of 81.7% at 6dB SNR of noisy speech.

  • PDF

Recognition of Emotion and Emotional Speech Based on Prosodic Processing

  • Kim, Sung-Ill
    • The Journal of the Acoustical Society of Korea
    • /
    • v.23 no.3E
    • /
    • pp.85-90
    • /
    • 2004
  • This paper presents two kinds of new approaches, one of which is concerned with recognition of emotional speech such as anger, happiness, normal, sadness, or surprise. The other is concerned with emotion recognition in speech. For the proposed speech recognition system handling human speech with emotional states, total nine kinds of prosodic features were first extracted and then given to prosodic identifier. In evaluation, the recognition results on emotional speech showed that the rates using proposed method increased more greatly than the existing speech recognizer. For recognition of emotion, on the other hands, four kinds of prosodic parameters such as pitch, energy, and their derivatives were proposed, that were then trained by discrete duration continuous hidden Markov models(DDCHMM) for recognition. In this approach, the emotional models were adapted by specific speaker's speech, using maximum a posteriori(MAP) estimation. In evaluation, the recognition results on emotional states showed that the rates on the vocal emotions gradually increased with an increase of adaptation sample number.

Effects of fractional fourier transform of facial images in face recognition using eigenfeatures (고유특징을 이용한 얼굴인식에 있어서 얼굴영상에 대한 분수차 Fourier 변환의 효과)

  • 심영미;장주석
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.35C no.8
    • /
    • pp.60-67
    • /
    • 1998
  • We studied the effects of fractional fourier transform in face recognition, in which only the amplitude spectra of transformed facial images were used.We used two recently developed face recognition methods, the most effective feature (MEF) method (i.e., eigenface method) and most discriminating feature (MDF) method, and the effects of th etransform for th etwo methods were consistent. We confirmed that the recognition rate by the use of MDF method is better than that consistent. We confirmed that the recognition rate by the use of MDF method is better than that by MEF regardless of the order to transform, these methods provided slightly better results when the order was 1 than for any other order values. Only when the order was close to 1, the recognition rates were robust to the shift of the input images, and the trend that the recognition rates decreased as the input size varied was independent of the order. From these results, we fond that it is most advantageous to use the amplitude spectra of the conventional fourier transform whose order is 1.

  • PDF

A Real-Time Implementation of Speech Recognition System Using Oak DSP core in the Car Noise Environment (자동차 환경에서 Oak DSP 코어 기반 음성 인식 시스템 실시간 구현)

  • Woo, K.H.;Yang, T.Y.;Lee, C.;Youn, D.H.;Cha, I.H.
    • Speech Sciences
    • /
    • v.6
    • /
    • pp.219-233
    • /
    • 1999
  • This paper presents a real-time implementation of a speaker independent speech recognition system based on a discrete hidden markov model(DHMM). This system is developed for a car navigation system to design on-chip VLSI system of speech recognition which is used by fixed point Oak DSP core of DSP GROUP LTD. We analyze recognition procedure with C language to implement fixed point real-time algorithms. Based on the analyses, we improve the algorithms which are possible to operate in real-time, and can verify the recognition result at the same time as speech ends, by processing all recognition routines within a frame. A car noise is the colored noise concentrated heavily on the low frequency segment under 400 Hz. For the noise robust processing, the high pass filtering and the liftering on the distance measure of feature vectors are applied to the recognition system. Recognition experiments on the twelve isolated command words were performed. The recognition rates of the baseline recognizer were 98.68% in a stopping situation and 80.7% in a running situation. Using the noise processing methods, the recognition rates were enhanced to 89.04% in a running situation.

  • PDF

Comparison of Speech Intelligibility & Performance of Speech Recognition in Real Driving Environments (자동차 주행 환경에서의 음성 전달 명료도와 음성 인식 성능 비교)

  • Lee Kwang-Hyun;Choi Dae-Lim;Kim Young-Il;Kim Bong-Wan;Lee Yong-Ju
    • MALSORI
    • /
    • no.50
    • /
    • pp.99-110
    • /
    • 2004
  • The normal transmission characteristics of sound are hardly obtained due to the various noises and structural factors in a running car environment. It is due to the channel distortion of the original source sound recorded by microphones, and it seriously degrades the performance of the speech recognition in real driving environments. In this paper we analyze the degree of intelligibility under the various sound distortion environments by channels according to driving speed with respect to speech transmission index(STI) and compare the STI with rates of speech recognition. We examine the correlation between measures of intelligibility depending on sound pick-up patterns and performance in speech recognition. Thereby we consider the optimal location of a microphone in single channel environment. In experimentation we find that high correlation is obtained between STI and rates of speech recognition.

  • PDF

A Study on the Submission of Multiple Candidates for Decision in Speaker-Independent Speech Recognition by VQ/HMM (VQ/HMM에 의한 화자독립 음성인식에서 다수 후보자를 인식 대상으로 제출하는 방법에 관한 연구)

  • Lee, Chang-Young;Nam, Ho-Soo
    • Speech Sciences
    • /
    • v.12 no.3
    • /
    • pp.115-124
    • /
    • 2005
  • We investigated on the submission of multiple candidates in speaker-independent speech recognition by VQ/HMM. Submission of fixed number of multiple candidates has first been examined. As the number of candidates increases by two, three, and four, the recognition error rates were found to decrease by 41%, 58%, and 65%, respectively compared to that of a single candidate. We tried another approach that the candidates within a range of Viterbi scores are submitted. The number of candidates showed geometric increase as the admitted range becomes large. For a practical application, a combination of the above two methods was also studied. We chose the candidates within some range of Viterbi scores and limited the maximum number of candidates submitted to five. Experimental results showed that recognition error rates of less than 10% could be achieved with average number of candidates of 3.2 by this method.

  • PDF