• Title/Summary/Keyword: 벡터화

Search Result 1,476, Processing Time 0.028 seconds

Efficiency Algorithm of Multispectral Image Compression in Wavelet Domain (웨이브릿 영역에서 다분광 화상 데이터의 효율적인 압축 알고리듬)

  • Park, Gyeong Nam;Kim, Yeong Chun;Jang, Jong Guk;Lee, Geon Il
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.38 no.4
    • /
    • pp.38-38
    • /
    • 2001
  • 본 논문에서는 웨이브릿 영역에서의 영역 분류와 대역간 예측 및 선택적 벡터 양자화를 이용한 다분광 화상테이타 압축 기법을 제안하였다. 이 방법에서는 각 대역을 웨이브릿 변환 후, 각 대역의 기저밴드의 대역별 특성을 이용하여 영역 분류를 행하였다. 그리고, 다른 대역과 해상도가 동일하고 공간적 분산이 작으며 분광적 상관성이 큰 기준대역 (reference channel)을 결정한 뒤, 이를 영역별 스칼라 및 분류별 가변 벡터 양자화를 행하여 부호화 하였다. 또한 기준대역과의 대역간 상관성이 큰 대역들에 대해서는 영역별 대역간 예측을 행한 후, 활동도가 높은 블록에 대해서만 선택적 벡터 양자화로 부호화를 행하였다. 이때, 활동도가 높은 블록들의 위치정보는 기준대역으로부터 얻어지는 임계치 지도 (threshold map; THMAP)를 이용하였다. 즉, 제안한 방법에서는 각 대역에 대해 웨이브릿 영역에서의 영역 분류 후 영역별 대역간 예측을 행함으로써 다분광 화상데이타에 존재하는 대역간 중복성을 제거하고 선택적 벡터 양자화를 행함으로써 대역내 중복성을 효과적으로 제거하여 압축효율을 향상시킨다. 실제 원격 센싱된 인공위성 화상데이타에 대한 실험을 통하여 제안한 기법의 부호화 효율이 기존의 기법에 비하여 우수함을 확인하였다.

Multispectral Image Compression Using Classified Interband Prediction and Vector Quantization in Wavelet domain (웨이브릿 영역에서의 영역별 대역간 예측과 벡터 양자화를 이용한 다분광 화상 데이타의 압축)

  • 반성원;권성근;이종원;박경남;김영춘;장종국;이건일
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.1B
    • /
    • pp.120-127
    • /
    • 2000
  • In this paper, we propose multispectral image compression using classified interband prediction and vector quantization in wavelet domain. This method classifies each region considering reflection characteristics of each band in image data. In wavelet domain, we perform the classified intraband VQ to remove intraband redundancy for a reference band image that has the lowest spatial variance and the best correlation with other band. And in wavelet domain, we perform the classifled interband prediction to remove interband redundancy for the remaining bands. Then error wavelet coefficients between original image and predicted image are intraband vector quantized to reduce prediction error. Experiments on remotely sensed satellite image show that coding efficiency of theproposed method is better than that of the conventional method.

  • PDF

XML Documents Clustering Technique Based on Bit Vector (비트벡터에 기반한 XML 문서 군집화 기법)

  • Kim, Woo-Saeng
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.47 no.5
    • /
    • pp.10-16
    • /
    • 2010
  • XML is increasingly important in data exchange and information management. A large amount of efforts have been spent in developing efficient techniques for accessing, querying, and storing XML documents. In this paper, we propose a new method to cluster XML documents efficiently. A bit vector which represents a XML document is proposed to cluster the XML documents. The similarity between two XML documents is measured by a bit-wise AND operation between two corresponding bit vectors. The experiment shows that the clusters are formed well and efficiently when a bit vector is used for the feature of a XML document.

Speaker Normalization using Gaussian Mixture Model for Speaker Independent Speech Recognition (화자독립 음성인식을 위한 GMM 기반 화자 정규화)

  • Shin, Ok-Keun
    • The KIPS Transactions:PartB
    • /
    • v.12B no.4 s.100
    • /
    • pp.437-442
    • /
    • 2005
  • For the purpose of speaker normalization in speaker independent speech recognition systems, experiments are conducted on a method based on Gaussian mixture model(GMM). The method, which is an improvement of the previous study based on vector quantizer, consists of modeling the probability distribution of canonical feature vectors by a GMM with an appropriate number of clusters, and of estimating the warp factor of a test speaker by making use of the obtained probabilistic model. The purpose of this study is twofold: improving the existing ML based methods, and comparing the performance of what is called 'soft decision' method with that of the previous study based on vector quantizer. The effectiveness of the proposed method is investigated by recognition experiments on the TIMIT corpus. The experimental results showed that a little improvement could be obtained tv adjusting the number of clusters in GMM appropriately.

Fast VQ Codebook Design by Sucessively Bisectioning of Principle Axis (주축의 연속적 분할을 통한 고속 벡터 양자화 코드북 설계)

  • Kang, Dae-Seong;Seo, Seok-Bae;Kim, Dai-Jin
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.4
    • /
    • pp.422-431
    • /
    • 2000
  • This paper proposes a new codebook generation method, called a PCA-Based VQ, that incorporates the PCA (Principal Component Analysis) technique into VQ (Vector Quantization) codebook design. The PCA technique reduces the data dimensions by transforming input image vectors into the feature vectors. The cluster of feature vectors in the transformed domain is bisectioned into two subclusters by an optimally chosen partitioning hyperplane. We expedite the searching of the optimal partitioning hyperplane that is the most time consuming process by considering that (1) the optimal partitioning hyperplane is perpendicular to the first principal axis of the feature vectors, (2) it is located on the equilibrium point of the left and right cluster's distortions, and (3) the left and right cluster's distortions can be adjusted incrementally. This principal axis bisectioning is successively performed on the cluster whose difference of distortion between before and after bisection is the maximum among the existing clusters until the total distortion of clusters becomes as small as the desired level. Simulation results show that the proposed PCA-based VQ method is promising because its reconstruction performance is as good as that of the SOFM (Self-Organizing Feature Maps) method and its codebook generation is as fast as that of the K-means method.

  • PDF

Vector Quantization Codebook Design Using Unbalanced Binary Tree and DCT Coefficients (불균형 이진트리와 DCT 계수를 이용한 벡터양자화 코드북)

  • 이경환;최정현;이법기;정원식;김경규;김덕규
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.12B
    • /
    • pp.2342-2348
    • /
    • 1999
  • DCT-based codebook design using binary tree was proposed to reduce computation time and to solve the initial codebook problem. In this method, DCT coefficient of training vectors that has maximum variance is to be a split key and the mean of coefficients at the location is used as split threshold, then balanced binary tree for final codebook is formed. However edge degradation appears in the reconstructed image, since the blocks of shade region are frequently selected for codevector. In this paper, we propose DCT-based vector quantization codebook design using unbalanced binary tree. Above all, the node that has the largest split key is splited. So the number of edge codevector can be increased. From the simulation results, this method reconstructs the edge region sincerely and shows higher PSNR than previous methods.

  • PDF

The Improved Binary Tree Vector Quantization Using Spatial Sensitivity of HVS (인간 시각 시스템의 공간 지각 특성을 이용한 개선된 이진트리 벡터양자화)

  • Ryu, Soung-Pil;Kwak, Nae-Joung;Ahn, Jae-Hyeong
    • The KIPS Transactions:PartB
    • /
    • v.11B no.1
    • /
    • pp.21-26
    • /
    • 2004
  • Color image quantization is a process of selecting a set of colors to display an image with some representative colors without noticeable perceived difference. It is very important in many applications to display a true color image in a low cost color monitor or printer. The basic problem is how to display 256 colors or less colors, called color palette, In this paper, we propose improved binary tree vector quantization based on spatial sensitivity which is one of the human visual properties. We combine the weights based on the responsibility of human visual system according to changes of three Primary colors in blocks of images with the process of splitting nodes using eigenvector in binary tree vector quantization. The test results show that the proposed method generates the quantized images with fine color and performs better than the conventional method in terms of clustering the similar regions. Also the proposed method can get the better result in subjective quality test and WSNR.

On the Lower Level Laplacian Pyramid Image Coding Using Vector Quantization (벡터 양자화를 이용한 저층 라플라시안 피라미드 영상의 부호화에 관한 연구)

  • 김정규;정호열;최태영
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.17 no.3
    • /
    • pp.213-224
    • /
    • 1992
  • An encoding technique based on region splitting and vector quantization is proposed for the lower level Laplacian pyramid images. The lower level Laplacian pyramid images have lower variance than higher levels but a great influence on compression ration due to large spatial area. And so from data compression viewpoint, we subdivide them with variance thresholding into two regions such as one called : flat region” and the other “edge region”, and encode the flat region with its mean value and the edge region as vector quantization method. The edge region can be reproduced faithfully and significant improvement on compression ratio can be accomplished with a little degradation of PSNR in spite of the effect of large flat region since the codebook used is generated from the edge region only on from the entire image including the flat region. It can be verified by computer simulation results that proposed method is more efficient in compression ratio and processing time than the conventional encoding technique of vector quantization.

  • PDF

The Image Compression Using the Central Vectors of Clusters (Cluster의 중심벡터를 이용하는 영상 압축)

  • Cho, Che-Hwang
    • The Journal of the Acoustical Society of Korea
    • /
    • v.14 no.1
    • /
    • pp.5-12
    • /
    • 1995
  • In the case where the set of training vectors constitute clusters, the codevectors of the codebook which is used to compression for speech and images in the vector quantization are regarded as the central vectors of the clusters constituted by given training vectors. In this work, we consider the distribution of Euclidean distance obtaining in the process of searching for the minimum distance between vectors, and propose the method searching for the proper number of and the central vectors of clusters. And then, the proposed method shows more than the about 4[dB] SNR than the LBG algorithm and the competitive learning algorithm

  • PDF

A fast block matching algorithm with adaptive search range (적응적 탐색범위를 사용한 블록정합 알고리듬)

  • 강문철;배황식;정정화
    • Proceedings of the IEEK Conference
    • /
    • 2003.07e
    • /
    • pp.1932-1935
    • /
    • 2003
  • 본 논문에서는 MPEG-2, MPEG-4, H.263 등에서 블록정합을 위해 사용되는 움직임 추정(Motion Estimation) 기법에서 적응적 탐색 범위를 기존의 알고리듬에 적용시킴으로써 계산량을 줄이고 화질도 개선하는 방법을 제안한다 제안된 알고리듬은 먼저 이웃한 움직임 벡터(Motion Vector)의 위치를 이용하여 예상된 움직임 벡터를 찾고 이 예상된 움직임 벡터의 X, Y 값의 크기를 작은 값, 중간 값, 큰 값, 세 가지로 분류해서 탐색범위를 적응적으로 변화시켜 움직임 벡터가 있을 확률이 큰 범위를 집중적으로 찾는다 그리고 각 분류에서 작은 값일 때는 전역 탐색을 적용하고 큰 값일 때는 기존의 알고리듬을 적용시키고 중간 값 일 때는 3단계탐색 기법을 적용시켜 더 적합한 움직임 벡터를 찾도록 하였다. 그리고 작은 값 일 때 구해진 움직임 벡터의 SAD(Sum of Absolute Difference) 값과 이웃한 움직임 벡터의 SAD값을 비교해 국소점에 빠졌다고 판단이 되면 다시 탐색 범위를 조정해서 움직임 벡터를 구함으로써 국소점에 빠지는 경우를 줄였다.

  • PDF