화자적응시스템을 위한 MLLR 알고리즘 연산량 감소

Reduction of Dimension of HMM parameters in MLLR Framework for Speaker Adaptation

  • 김지운 (인하대학교 전자공학과 DSP Lab.) ;
  • 정재호 (인하대학교 전자공학과 DSP Lab.)
  • 발행 : 2003.05.01

초록

We discuss how to reduce the number of inverse matrix and its dimensions requested in MLLR framework for speaker adaptation. To find a smaller set of variables with less redundancy, we employ PCA(principal component analysis) and ICA(independent component analysis) that would give as good a representation as possible. The amount of additional computation when PCA or ICA is applied is as small as it can be disregarded. The dimension of HMM parameters is reduced to about 1/3 ~ 2/7 dimensions of SI(speaker independent) model parameter with which speech recognition system represents word recognition rate as much as ordinary MLLR framework. If dimension of SI model parameter is n, the amount of computation of inverse matrix in MLLR is proportioned to O($n^4$). So, compared with ordinary MLLR, the amount of total computation requested in speaker adaptation is reduced to about 1/80~1/150.

키워드