• Title/Summary/Keyword: Compression Algorithm

Search Result 1,096, Processing Time 0.024 seconds

Design of the Compression Algorithm for in-Memory Data of the Virtual Memory (가상 메모리 압축을 위한 CAMD 알고리즘 설계)

  • Jang, Seung-Ju
    • The KIPS Transactions:PartA
    • /
    • v.11A no.3
    • /
    • pp.157-162
    • /
    • 2004
  • This paper suggests the CAMD(Compression Algorithm for in-Memory Data) algorithm that is not moved the pages into the swap space by assigning the compressed cache area in the main memory. The CAMD algorithm that supports the virtual memory system takes high memory usability and performance benefit by reducing the page fault. The memory data is not general data. It is extraordinary data format. In general it consists of specific form of data. Therefore. the CAMD algorithm can compress this data efficiently.

Adaptive Rank-Reindexing Algorithm for Lossless Index Image Compression (무손실 인덱스 영상 압축을 위한 적응적 랭크-리인덱싱 알고리즘)

  • Lee Han-Jeong;Yoo Gi-Hyung;Kim Hyung-Moo;You Kang-Soo;Kwak Hoon-Sung
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.54 no.8
    • /
    • pp.501-503
    • /
    • 2005
  • In this paper, using ranks of co-occurrence frequency about indices in pairs of neighboring pixels, we introduce a new re-indexing algorithm for efficiency of index color image lossless compression. The proposed algorithm is suitable for arithmetic coding because it has concentrated distributions of small variance. Experimental results proved that the proposed algorithm reduces the bit rates than other coding schemes, more specifically $15\%$, $54\%$ and $12\%$ for LZW algorithm of GIF, the plain arithmetic coding method and Zeng's scheme, respectively.

Low Power Scan Chain Reordering Method with Limited Routing Congestion for Code-based Test Data Compression

  • Kim, Dooyoung;Ansari, M. Adil;Jung, Jihun;Park, Sungju
    • JSTS:Journal of Semiconductor Technology and Science
    • /
    • v.16 no.5
    • /
    • pp.582-594
    • /
    • 2016
  • Various test data compression techniques have been developed to reduce the test costs of system-on-a-chips. In this paper, a scan chain reordering algorithm for code-based test data compression techniques is proposed. Scan cells within an acceptable relocation distance are ranked to reduce the number of conflicts in all test patterns and rearranged by a positioning algorithm to minimize the routing overhead. The proposed method is demonstrated on ISCAS '89 benchmark circuits with their physical layout by using a 180 nm CMOS process library. Significant improvements are observed in compression ratio and test power consumption with minor routing overhead.

A New Proposal of Adaptive BTC for Image Data Compression (畵像壓縮을 위한 適應 BTC 方法의 提案)

  • Jang, Ki-Soong;Oh, Seong-Mock;Lee, Young-Choul
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.26 no.7
    • /
    • pp.125-131
    • /
    • 1989
  • This paper proposes a new ABTC (Adaptive Block Truncation Coding) algorithm which improves the BTC algorithm for image data compression. A new adaptive block truncation coding which adopts a selective coding scheme depending on the local characteristics of an image has been described. The characteristics of the ABTC algorithm can be summarized as high compression ratio and the algorithm simplicity. Using this algorithm, color images can be coded at a variable bit rate from 1.0 (bit/pel) to 2.56 (bit/pel) and high compression rate (1.3-105 bit/pel) can be achieved without conspicuous image degradation compared with original images.

  • PDF

The Variable Block-based Image Compression Technique using Wavelet Transform (웨이블릿 변환을 이용한 가변블록 기반 영상 압축)

  • 권세안;장우영;송광훈
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.7B
    • /
    • pp.1378-1383
    • /
    • 1999
  • In this paper, an effective variable-block-based image compression technique using wavelet transform is proposed. Since the statistical property of each wavelet subband is different, we apply the adaptive quantization to each wavelet subband. In the proposed algorithm, each subband is divided into non-overlapping variable-sized blocks based on directional properties. In addition, we remove wavelet coefficients which are below a certain threshold value for coding efficiency. To compress the transformed data, the proposed algorithm quantizes the wavelet coefficients using scalar quantizer in LL subband and vector quantizers for other subbands to increase compression ratio. The proposed algorithm shows improvements in compression ratio as well as PSNR compared with the existing block-based compression algorithms. In addition, it does not cause any blocking artifacts in very low bit rates even though it is also a block-based method. The proposed algorithm also has advantage in computational complexity over the existing wavelet-based compression algorithms since it is a block-based algorithm.

  • PDF

A New Predictive EC Algorithm for Reduction of Memory Size and Bandwidth Requirements in Wavelet Transform (웨이블릿 변환의 메모리 크기와 대역폭 감소를 위한 Prediction 기반의 Embedded Compression 알고리즘)

  • Choi, Woo-Soo;Son, Chang-Hoon;Kim, Ji-Won;Na, Seong-Yu;Kim, Young-Min
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.7
    • /
    • pp.917-923
    • /
    • 2011
  • In this paper, a new prediction based embedded compression (EC) codec algorithm for the JPEG2000 encoder system is proposed to reduce excessive memory requirements. The EC technique can reduce the 50 % memory requirement for intermediate low-frequency coefficients during multiple discrete wavelet transform (DWT) stages compared with direct implementation of the DWT engine of this paper. The LOCO-I predictor and MAP are widely used in many lossless picture compression codec. The proposed EC algorithm use these predictor which are very simple but surprisingly effective. The predictive EC scheme adopts a forward adaptive quantization and fixed length coding to encoding the prediction error. Simulation results show that our LOCO-I and MAP based EC codecs present only PSNR degradation of 0.48 and 0.26 dB in average, respectively. The proposed algorithm improves the average PSNR by 1.39 dB compared to the previous work in [9].

Embedded Compression Codec Algorithm for Motion Compensated Wavelet Video Coding System (움직임 보상된 웨이블릿 기반의 비디오 코딩 시스템에 적용 가능한 임베디드 압축 코덱 알고리즘)

  • Kim, Song-Ju
    • The Journal of the Korea Contents Association
    • /
    • v.12 no.3
    • /
    • pp.77-83
    • /
    • 2012
  • In this paper, a low-complexity embedded compression (EC) Codec algorithm for the wavelet video coder is applied to reduce excessive external memory requirements. The EC algorithm is used to achieve a fixed compression ratio of 50 % under the near-lossless-compression constraint. The EC technique can reduce the 50 % memory requirement for intermediate low-frequency coefficients during multiple discrete wavelet transform stages compared with direct implementation of the wavelet video encoder of this paper. Furthermore, the EC scheme based on a forward adaptive quantization and fixed length coding can save bandwidth and size of buffer between DWT and SPIHT to 50 %. Simulation results show that our EC algorithm present only PSNR degradation of 0.179 and 0.162 dB in average when the target bit-rate of the video coder are 1 and 0.5 bpp, respectively.

Proposal for Decoding-Compatible Parallel Deflate Algorithm by Inserting Control Header Composed of Non-Compressed Blocks (비 압축 블록으로 구성된 제어 헤더 삽입을 통한 압축 해제 호환성 있는 병렬 처리 Deflate 알고리즘 제안)

  • Kim Jung Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.5
    • /
    • pp.207-216
    • /
    • 2023
  • For decoding-compatible parallel Deflate algorithm, this study proposed a new method of the control header being made in such a way that essential information for parallel compression and decompression are stored in the Disposed Bit Area (DBA) of the non-compression block and being inserted into the compressed blocks. Through this, parallel compression and decompression are possible while maintaining perfect compatibility with the existing decoder. After applying this method, the compression time was reduced by up to 71.2% compared to the sequential processing method, and the parallel decompression time was reduced by up to 65.7%. In particular, it is well known that parallel decompression is impossible due to the structural limitations of the Deflate algorithm. However, the decoder equipped with the proposed method enables high-speed parallel decompression at the algorithm level and maintains compatibility, so that parallelly compressed data can be decoded normally by existing decoder programs.

Implementation of Real Time 3 channel Transmission System Using ECG Data Compression Algorithm by Max-Min Slope Update (최대 및 최소 기울기 갱신에 의한 ECG 압축 알고리듬을 이용한 실시간 3채널 전송시스템 구현)

  • 조진호;김명남
    • Journal of Biomedical Engineering Research
    • /
    • v.16 no.3
    • /
    • pp.271-278
    • /
    • 1995
  • An ECG data compression algorithM using max-min slope update is proposed and a real time 3 channel ECG transmission system is implemented using the proposed algorithm. In order to effectively compress ECG data, we compare a threshold value with the max-min slope difference (MMSD) which is updated at each sample values. If this MMSD value is smaller than the threshold value, then the data is compressed. Conversely, when the MMSD value is larger than threshold value, the data is transmitted after storing the value and the length between the data which is beyond previous threshold level. As a result, it can accurately compress both the region of QRS, P, and T wave that has fast-changing and the region of the base line that slope is changing slow. Therefore, it Is possible to enhance the compression rate and the percent roms difference. In addition, because of the simplicity, this algorithm is more suitable for real-time implementation.

  • PDF

A Hybrid Hardware Architecture for LCD Overdrive Frame Buffer Reduction

  • Choi, Ji-Yong;Jeong, An-Sun;Baek, Jun-Geun
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 2007.08a
    • /
    • pp.814-817
    • /
    • 2007
  • We present a hybrid hardware architecture capable of encoding and decoding a full HD resolution video with 60 fps. A number of technical modifications are applied to an existing image compression algorithm not only to accommodate large videos at a high frame rate but to enhance the quality of synthetic images, such as characters and video game images. Image quality of the proposed algorithm at a 1/6 compression ratio was comparable to that of the BTC based 1/3 compression algorithm.

  • PDF