• Title/Summary/Keyword: recognition-rate

Search Result 2,809, Processing Time 0.03 seconds

Development of RFID Management System for Packaged Liquid Food Logistics (I) - Analysis of RFID Recognition Performance by Level of Water - (용기포장 액상 식품의 물류관리를 위한 RFID 시스템 개발(I) - 물의 높이에 따른 RFID 인식성능 분석 -)

  • Kim, Yong-Joo;Kim, Tae-Hyeong
    • Journal of Biosystems Engineering
    • /
    • v.34 no.6
    • /
    • pp.454-461
    • /
    • 2009
  • The purpose of this study is to analyze the RFID recognition performance by level of water. A 13.56 MHz RFID management system for packaged liquid food logistics is consisted of antenna, reader, passive type tags, and embedded controller. The tests were conducted at different level of water, distances between tag and antenna, and position of attached tags. To analyze the RFID recognition performance, maximum recognition distances for a container and recognition rates for a logistics made of 27 containers were measured and analyzed. The maximum recognition distance for a container was different depending on position of attached tags, and attached tag at upside position showed a good performance. But, the recognition rate of 27 containers showed a good ability for attached tags at front side position, 30~35 cm distance to antenna, and water level 1. Therefore, to manage packaged liquid food logistics using RFID system, position of attached tag, distances between tag and antenna, and level of water should be considered.

A Study on Korean 4-connected Digit Recognition Using Demi-syllable Context-dependent Models (반음절 문맥종속 모델을 이용한 한국어 4 연숫자음 인식에 관한 연구)

  • 이기영;최성호;이호영;배명진
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.3
    • /
    • pp.175-181
    • /
    • 2003
  • Because a word of Korean digits is a syllable and deeply coarticulatied in connected digits, some recognition models based on demisyllables have been proposed by researchers. However, they could not show an excellent recognition results yet. This paper proposes a recognition model based on extended and context-dependent demisyllables, such as a tri-demisyllable like a tri-phone, for the Korean 4-connected digits recognition. For experiments, we use a toolkit of HTK 3.0 for building this model of continuous HMMs using training Korean connected digits from SiTEC database and for recognizing unknown ones. The results show that the recognition rate is 92% and this model has an ability to improve the recognition performance of Korean connected digits.

Design of Face Recognition System Based on Pose Estimation : Comparative Studies of Pose Estimation Algorithms (포즈 추정 기반 얼굴 인식 시스템 설계 : 포즈 추정 알고리즘 비교 연구)

  • Kim, Jin-Yul;Kim, Jong-Bum;Oh, Sung-Kwun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.66 no.4
    • /
    • pp.672-681
    • /
    • 2017
  • This paper is concerned with the design methodology of face recognition system based on pose estimation. In 2-dimensional face recognition, the variations of facial pose cause the deterioration of recognition performance because object recognition is carried out by using brightness of each pixel on image. To alleviate such problem, the proposed face recognition system deals with Learning Vector Quantizatioin(LVQ) or K-Nearest Neighbor(K-NN) to estimate facial pose on image and then the images obtained from LVQ or K-NN are used as the inputs of networks such as Convolution Neural Networks(CNNs) and Radial Basis Function Neural Networks(RBFNNs). The effectiveness and efficiency of the post estimation using LVQ and K-NN as well as face recognition rate using CNNs and RBFNNs are discussed through experiments carried out by using ICPR and CMU PIE databases.

Speaker-dependent Speech Recognition Algorithm for Male and Female Classification (남녀성별 분류를 위한 화자종속 음성인식 알고리즘)

  • Choi, Jae-Seung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.17 no.4
    • /
    • pp.775-780
    • /
    • 2013
  • This paper proposes a speaker-dependent speech recognition algorithm which can classify the gender for male and female speakers in white noise and car noise, using a neural network. The proposed speech recognition algorithm is trained by the neural network to recognize the gender for male and female speakers, using LPC (Linear Predictive Coding) cepstrum coefficients. In the experiment results, the maximal improvement of total speech recognition rate is 96% for white noise and 88% for car noise, respectively, after trained a total of six neural networks. Finally, the proposed speech recognition algorithm is compared with the results of a conventional speech recognition algorithm in the background noisy environment.

Improved Two-Phase Framework for Facial Emotion Recognition

  • Yoon, Hyunjin;Park, Sangwook;Lee, Yongkwi;Han, Mikyong;Jang, Jong-Hyun
    • ETRI Journal
    • /
    • v.37 no.6
    • /
    • pp.1199-1210
    • /
    • 2015
  • Automatic emotion recognition based on facial cues, such as facial action units (AUs), has received huge attention in the last decade due to its wide variety of applications. Current computer-based automated two-phase facial emotion recognition procedures first detect AUs from input images and then infer target emotions from the detected AUs. However, more robust AU detection and AU-to-emotion mapping methods are required to deal with the error accumulation problem inherent in the multiphase scheme. Motivated by our key observation that a single AU detector does not perform equally well for all AUs, we propose a novel two-phase facial emotion recognition framework, where the presence of AUs is detected by group decisions of multiple AU detectors and a target emotion is inferred from the combined AU detection decisions. Our emotion recognition framework consists of three major components - multiple AU detection, AU detection fusion, and AU-to-emotion mapping. The experimental results on two real-world face databases demonstrate an improved performance over the previous two-phase method using a single AU detector in terms of both AU detection accuracy and correct emotion recognition rate.

Emotion Robust Speech Recognition using Speech Transformation (음성 변환을 사용한 감정 변화에 강인한 음성 인식)

  • Kim, Weon-Goo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.20 no.5
    • /
    • pp.683-687
    • /
    • 2010
  • This paper studied some methods which use frequency warping method that is the one of the speech transformation method to develope the robust speech recognition system for the emotional variation. For this purpose, the effect of emotional variations on the speech signal were studied using speech database containing various emotions and it is observed that speech spectrum is affected by the emotional variation and this effect is one of the reasons that makes the performance of the speech recognition system worse. In this paper, new training method that uses frequency warping in training process is presented to reduce the effect of emotional variation and the speech recognition system based on vocal tract length normalization method is developed to be compared with proposed system. Experimental results from the isolated word recognition using HMM showed that new training method reduced the error rate of the conventional recognition system using speech signal containing various emotions.

2-D Conditional Moment for Recognition of Deformed Letters

  • Yoon, Myoong-Young
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.6 no.2
    • /
    • pp.16-22
    • /
    • 2001
  • In this paper we mose a new scheme for recognition of deformed letters by extracting feature vectors based on Gibbs distributions which are well suited for representing the spatial continuity. The extracted feature vectors are comprised of 2-D conditional moments which are invariant under translation, rotation, and scale of an image. The Algorithm for pattern recognition of deformed letters contains two parts: the extraction of feature vector and the recognition process. (i) We extract feature vector which consists of an improved 2-D conditional moments on the basis of estimated conditional Gibbs distribution for an image. (ii) In the recognition phase, the minimization of the discrimination cost function for a deformed letters determines the corresponding template pattern. In order to evaluate the performance of the proposed scheme, recognition experiments with a generated document was conducted. on Workstation. Experiment results reveal that the proposed scheme has high recognition rate over 96%.

  • PDF

Online Recognition of Handwritten Korean and English Characters

  • Ma, Ming;Park, Dong-Won;Kim, Soo Kyun;An, Syungog
    • Journal of Information Processing Systems
    • /
    • v.8 no.4
    • /
    • pp.653-668
    • /
    • 2012
  • In this study, an improved HMM based recognition model is proposed for online English and Korean handwritten characters. The pattern elements of the handwriting model are sub character strokes and ligatures. To deal with the problem of handwriting style variations, a modified Hierarchical Clustering approach is introduced to partition different writing styles into several classes. For each of the English letters and each primitive grapheme in Korean characters, one HMM that models the temporal and spatial variability of the handwriting is constructed based on each class. Then the HMMs of Korean graphemes are concatenated to form the Korean character models. The recognition of handwritten characters is implemented by a modified level building algorithm, which incorporates the Korean character combination rules within the efficient network search procedure. Due to the limitation of the HMM based method, a post-processing procedure that takes the global and structural features into account is proposed. Experiments showed that the proposed recognition system achieved a high writer independent recognition rate on unconstrained samples of both English and Korean characters. The comparison with other schemes of HMM-based recognition was also performed to evaluate the system.

A MFCC-based CELP Speech Coder for Server-based Speech Recognition in Network Environments (네트워크 환경에서 서버용 음성 인식을 위한 MFCC 기반 음성 부호화기 설계)

  • Lee, Gil-Ho;Yoon, Jae-Sam;Oh, Yoo-Rhee;Kim, Hong-Kook
    • MALSORI
    • /
    • no.54
    • /
    • pp.27-43
    • /
    • 2005
  • Existing standard speech coders can provide speech communication of high quality while they degrade the performance of speech recognition systems that use the reconstructed speech by the coders. The main cause of the degradation is that the spectral envelope parameters in speech coding are optimized to speech quality rather than to the performance of speech recognition. For example, mel-frequency cepstral coefficient (MFCC) is generally known to provide better speech recognition performance than linear prediction coefficient (LPC) that is a typical parameter set in speech coding. In this paper, we propose a speech coder using MFCC instead of LPC to improve the performance of a server-based speech recognition system in network environments. However, the main drawback of using MFCC is to develop the efficient MFCC quantization with a low-bit rate. First, we explore the interframe correlation of MFCCs, which results in the predictive quantization of MFCC. Second, a safety-net scheme is proposed to make the MFCC-based speech coder robust to channel error. As a result, we propose a 8.7 kbps MFCC-based CELP coder. It is shown from a PESQ test that the proposed speech coder has a comparable speech quality to 8 kbps G.729 while it is shown that the performance of speech recognition using the proposed speech coder is better than that using G.729.

  • PDF

Multi-classifier Fusion Based Facial Expression Recognition Approach

  • Jia, Xibin;Zhang, Yanhua;Powers, David;Ali, Humayra Binte
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.1
    • /
    • pp.196-212
    • /
    • 2014
  • Facial expression recognition is an important part in emotional interaction between human and machine. This paper proposes a facial expression recognition approach based on multi-classifier fusion with stacking algorithm. The kappa-error diagram is employed in base-level classifiers selection, which gains insights about which individual classifier has the better recognition performance and how diverse among them to help improve the recognition accuracy rate by fusing the complementary functions. In order to avoid the influence of the chance factor caused by guessing in algorithm evaluation and get more reliable awareness of algorithm performance, kappa and informedness besides accuracy are utilized as measure criteria in the comparison experiments. To verify the effectiveness of our approach, two public databases are used in the experiments. The experiment results show that compared with individual classifier and two other typical ensemble methods, our proposed stacked ensemble system does recognize facial expression more accurately with less standard deviation. It overcomes the individual classifier's bias and achieves more reliable recognition results.