• Title/Summary/Keyword: Baum-Welch algorithm

Search Result 19, Processing Time 0.037 seconds

HMM-Based Bandwidth Extension Using Baum-Welch Re-Estimation Algorithm (Baum-Welch 학습법을 이용한 HMM 기반 대역폭 확장법)

  • Song, Geun-Bae;Kim, Austin
    • The Journal of the Acoustical Society of Korea
    • /
    • v.26 no.6
    • /
    • pp.259-268
    • /
    • 2007
  • This paper contributes to an improvement of the statistical bandwidth extension(BWE) system based on Hidden Markov Model(HMM). First, the existing HMM training method for BWE, which is suggested originally by Jax, is analyzed in comparison with the general Baum-Welch training method. Next, based on this analysis, a new HMM-based BWE method is suggested which adopts the Baum-Welch re-estimation algorithm instead of the Jax's to train HMM model. Conclusionally speaking, the Baum-Welch re-estimation algorithm is a generalized form of the Jax's training method. It is flexible and adaptive in modeling the statistical characteristic of training data. Therefore, it generates a better model to the training data, which results in an enhanced BWE system. According to experimental results, the new method performs much better than the Jax's BWE systemin all cases. Under the given test conditions, the RMS log spectral distortion(LSD) scores were improved ranged from 0.31dB to 0.8dB, and 0.52dB in average.

Improved Automatic Lipreading by Multiobjective Optimization of Hidden Markov Models (은닉 마르코프 모델의 다목적함수 최적화를 통한 자동 독순의 성능 향상)

  • Lee, Jong-Seok;Park, Cheol-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.15B no.1
    • /
    • pp.53-60
    • /
    • 2008
  • This paper proposes a new multiobjective optimization method for discriminative training of hidden Markov models (HMMs) used as the recognizer for automatic lipreading. While the conventional Baum-Welch algorithm for training HMMs aims at maximizing the probability of the data of a class from the corresponding HMM, we define a new training criterion composed of two minimization objectives and develop a global optimization method of the criterion based on simulated annealing. The result of a speaker-dependent recognition experiment shows that the proposed method improves performance by the relative error reduction rate of about 8% in comparison to the Baum-Welch algorithm.

Hidden Markov Model-based Extraction of Internet Information (은닉 마코브 모델을 이용한 인터넷 정보 추출)

  • Park, Dong-Chul
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.46 no.3
    • /
    • pp.8-14
    • /
    • 2009
  • A Hidden Markov Model(HMM)-based information extraction method is proposed in this paper. The proposed extraction method is applied to extraction of products' prices. The input of the proposed IESHMM is the URLs of a search engine's interface, which contains the names of the product types. The output of the system is the list of extracted slots of each product: name, price, image, and URL. With the observation data set Maximum Likelihood algorithm and Baum-Welch algorithm are used for the training of HMM and The Viterbi algorithm is then applied to find the state sequence of the maximal probability that matches the observation block sequence. When applied to practical problems, the proposed HMM-based system shows improved results over a conventional method, PEWEB, in terms of recall ration and accuracy.

A Data-Driven Jacobian Adaptation Method for the Noisy Speech Recognition (잡음음성인식을 위한 데이터 기반의 Jacobian 적응방식)

  • Chung Young-Joo
    • The Journal of the Acoustical Society of Korea
    • /
    • v.25 no.4
    • /
    • pp.159-163
    • /
    • 2006
  • In this paper a data-driven method to improve the performance of the Jacobian adaptation (JA) for the noisy speech recognition is proposed. In stead of constructing the reference HMM by using the model composition method like the parallel model combination (PMC), we propose to train the reference HMM directly with the noisy speech. This was motivated from the idea that the directly trained reference HMM will model the acoustical variations due to the noise better than the composite HMM. For the estimation of the Jacobian matrices, the Baum-Welch algorithm is employed during the training. The recognition experiments have been done to show the improved performance of the proposed method over the Jacobian adaptation as well as other model compensation methods.

A Recognition Algorithm of Suspicious Human Behaviors using Hidden Markov Models in an Intelligent Surveillance System (지능형 영상 감시 시스템에서의 은닉 마르코프 모델을 이용한 특이 행동 인식 알고리즘)

  • Jung, Chang-Wook;Kang, Dong-Joong
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.11
    • /
    • pp.1491-1500
    • /
    • 2008
  • This paper proposes an intelligent surveillance system to recognize suspicious patterns of the human behavior by using the Hidden Markov Model. First, the method finds foot area of the human by motion detection algorithm from image sequence of the surveillance camera. Then, these foot locus form observation series of features to learn the HMM. The feature that is position of the human foot is changed to each code that corresponds to a specific label among 16 local partitions of image region. Therefore, specific moving patterns formed by the foot locus are the series of the label numbers. The Baum-Welch algorithm of the HMM learns each suspicious and specific pattern to classify the human behaviors. To recognize the inputted human behavior pattern in a test image, the probabilistic comparison between the learned pattern of the HMM and foot series to be tested decides the categorization of the test pattern. The experimental results show that the method can be applied to detect a suspicious person prowling in corridor.

  • PDF

Performance Comparison of GMM and HMM Approaches for Bandwidth Extension of Speech Signals (음성신호의 대역폭 확장을 위한 GMM 방법 및 HMM 방법의 성능평가)

  • Song, Geun-Bae;Kim, Austin
    • The Journal of the Acoustical Society of Korea
    • /
    • v.27 no.3
    • /
    • pp.119-128
    • /
    • 2008
  • This paper analyzes the relationship between two representative statistical methods for bandwidth extension (BWE): Gaussian Mixture Model (GMM) and Hidden Markov Model (HMM) ones, and compares their performances. The HMM method is a memory-based system which was developed to take advantage of the inter-frame dependency of speech signals. Therefore, it could be expected to estimate better the transitional information of the original spectra from frame to frame. To verify it, a dynamic measure that is an approximation of the 1st-order derivative of spectral function over time was introduced in addition to a static measure. The comparison result shows that the two methods are similar in the static measure, while, in the dynamic measure, the HMM method outperforms explicitly the GMM one. Moreover, this difference increases in proportion to the number of states of HMM model. This indicates that the HMM method would be more appropriate at least for the 'blind BWE' problem. On the other hand, nevertheless, the GMM method could be treated as a preferable alternative of the HMM one in some applications where the static performance and algorithm complexity are critical.

A Real-Time Embedded Speech Recognition System (실시간 임베디드 음성 인식 시스템)

  • 남상엽;전은희;박인정
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.40 no.1
    • /
    • pp.74-81
    • /
    • 2003
  • In this study, we'd implemented a real time embedded speech recognition system that requires minimum memory size for speech recognition engine and DB. The word to be recognized consist of 40 commands used in a PCS phone and 10 digits. The speech data spoken by 15 male and 15 female speakers was recorded and analyzed by short time analysis method, which window size is 256. The LPC parameters of each frame were computed through Levinson-Burbin algorithm and they were transformed to Cepstrum parameters. Before the analysis, speech data should be processed by pre-emphasis that will remove the DC component in speech and emphasize high frequency band. Baum-Welch reestimation algorithm was used for the training of HMM. In test phone, we could get a recognition rate using likelihood method. We implemented an embedded system by porting the speech recognition engine on ARM core evaluation board. The overall recognition rate of this system was 95%, while the rate on 40 commands was 96% and that 10 digits was 94%.

Discrete HMM Training Algorithm for Incomplete Time Series Data (불완전 시계열 데이터를 위한 이산 HMM 학습 알고리듬)

  • Sin, Bong-Kee
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.1
    • /
    • pp.22-29
    • /
    • 2016
  • Hidden Markov Model is one of the most successful and popular tools for modeling real world sequential data. Real world signals come in a variety of shapes and variabilities, among which temporal and spectral ones are the prime targets that the HMM aims at. A new problem that is gaining increasing attention is characterizing missing observations in incomplete data sequences. They are incomplete in that there are holes or omitted measurements. The standard HMM algorithms have been developed for complete data with a measurements at each regular point in time. This paper presents a modified algorithm for a discrete HMM that allows substantial amount of omissions in the input sequence. Basically it is a variant of Baum-Welch which explicitly considers the case of isolated or a number of omissions in succession. The algorithm has been tested on online handwriting samples expressed in direction codes. An extensive set of experiments show that the HMM so modeled are highly flexible showing a consistent and robust performance regardless of the amount of omissions.

Content-based Image Retrieval using an Improved Chain Code and Hidden Markov Model (개선된 chain code와 HMM을 이용한 내용기반 영상검색)

  • 조완현;이승희;박순영;박종현
    • Proceedings of the IEEK Conference
    • /
    • 2000.09a
    • /
    • pp.375-378
    • /
    • 2000
  • In this paper, we propose a novo] content-based image retrieval system using both Hidden Markov Model(HMM) and an improved chain code. The Gaussian Mixture Model(GMM) is applied to statistically model a color information of the image, and Deterministic Annealing EM(DAEM) algorithm is employed to estimate the parameters of GMM. This result is used to segment the given image. We use an improved chain code, which is invariant to rotation, translation and scale, to extract the feature vectors of the shape for each image in the database. These are stored together in the database with each HMM whose parameters (A, B, $\pi$) are estimated by Baum-Welch algorithm. With respect to feature vector obtained in the same way from the query image, a occurring probability of each image is computed by using the forward algorithm of HMM. We use these probabilities for the image retrieval and present the highest similarity images based on these probabilities.

  • PDF

Training HMM Structure and Parameters with Genetic Algorithm and Harmony Search Algorithm

  • Ko, Kwang-Eun;Park, Seung-Min;Park, Jun-Heong;Sim, Kwee-Bo
    • Journal of Electrical Engineering and Technology
    • /
    • v.7 no.1
    • /
    • pp.109-114
    • /
    • 2012
  • In this paper, we utilize training strategy of hidden Markov model (HMM) to use in versatile issues such as classification of time-series sequential data such as electric transient disturbance problem in power system. For this, an automatic means of optimizing HMMs would be highly desirable, but it raises important issues: model interpretation and complexity control. With this in mind, we explore the possibility of using genetic algorithm (GA) and harmony search (HS) algorithm for optimizing the HMM. GA is flexible to allow incorporating other methods, such as Baum-Welch, within their cycle. Furthermore, operators that alter the structure of HMMs can be designed to simple structures. HS algorithm with parameter-setting free technique is proper for optimizing the parameters of HMM. HS algorithm is flexible so as to allow the elimination of requiring tedious parameter assigning efforts. In this paper, a sequential data analysis simulation is illustrated, and the optimized-HMMs are evaluated. The optimized HMM was capable of classifying a sequential data set for testing compared with the normal HMM.