• Title/Summary/Keyword: lossless data compression

Search Result 73, Processing Time 0.021 seconds

Adaptive Prediction for Lossless Image Compression

  • Park, Sang-Ho
    • Proceedings of the Korea Society of Information Technology Applications Conference
    • /
    • 2005.11a
    • /
    • pp.169-172
    • /
    • 2005
  • Genetic algorithm based predictor for lossless image compression is propsed. We describe a genetic algorithm to learn predictive model for lossless image compression. The error image can be further compressed using entropy coding such as Huffman coding or arithmetic coding. We show that the proposed algorithm can be feasible to lossless image compression algorithm.

  • PDF

A New Method of Lossless Universal Data Compression (새로운 무손실 유니버셜 데이터 압축 기법)

  • Kim, Sung-Soo;Lee, Hae-Kee
    • The Transactions of the Korean Institute of Electrical Engineers P
    • /
    • v.58 no.3
    • /
    • pp.285-290
    • /
    • 2009
  • In this paper, we propose a new algorithm that improves the lossless data compression rate. The proposed algorithm lessens the redundancy and improves the compression rate evolutionarily around 40 up to 80 percentile depending on the characteristics of binary images used for compression. In order to demonstrate the superiority of the proposed method, the comparison between the proposed method and the LZ78 (LZ77) is demonstrated through experimental results theoretical analysis.

Lossless Time Series Data Compression Technique Applicable to Edge Nodes (에지노드에 적용 가능한 무손실 시계열데이터 압축 기법)

  • Soosung Lee;Sang-Ho Hwang;Sungho Kim;Jang-Kyu Yun;Yong-Wan Park
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.19 no.4
    • /
    • pp.195-201
    • /
    • 2024
  • In this paper, we propose an improved technique called HAB (Huffman encoding Aware Bitpacking) that enhances the Sprintz method, which is a representative time series data compression technique. The proposed technique boosts compression rates by incorporating Huffman encoding-aware bitpacking in the secondary compression stage. Additionally, it can be applied to terminal nodes or edge nodes with limited available resources, as it does not require separate parameters or storage space for compression. The proposed technique is a lossless method and is suitable for fields that require the generation of artificial intelligence models and accurate data analysis. In the experimental evaluation, the proposed HAB showed an average improvement of 14.7% compared to the existing technique in terms of compression rate.

A Preprocessing Algorithm for Efficient Lossless Compression of Gray Scale Images

  • Kim, Sun-Ja;Hwang, Doh-Yeun;Yoo, Gi-Hyoung;You, Kang-Soo;Kwak, Hoon-Sung
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.2485-2489
    • /
    • 2005
  • This paper introduces a new preprocessing scheme to replace original data of gray scale images with particular ordered data so that performance of lossless compression can be improved more efficiently. As a kind of preprocessing technique to maximize performance of entropy encoder, the proposed method converts the input image data into more compressible form. Before encoding a stream of the input image, the proposed preprocessor counts co-occurrence frequencies for neighboring pixel pairs. Then, it replaces each pair of adjacent gray values with particular ordered numbers based on the investigated co-occurrence frequencies. When compressing ordered image using entropy encoder, we can expect to raise compression rate more highly because of enhanced statistical feature of the input image. In this paper, we show that lossless compression rate increased by up to 37.85% when comparing results from compressing preprocessed and non-preprocessed image data using entropy encoder such as Huffman, Arithmetic encoder.

  • PDF

An Efficient Bit-Level Lossless Grayscale Image Compression Based on Adaptive Source Mapping

  • Al-Dmour, Ayman;Abuhelaleh, Mohammed;Musa, Ahmed;Al-Shalabi, Hasan
    • Journal of Information Processing Systems
    • /
    • v.12 no.2
    • /
    • pp.322-331
    • /
    • 2016
  • Image compression is an essential technique for saving time and storage space for the gigantic amount of data generated by images. This paper introduces an adaptive source-mapping scheme that greatly improves bit-level lossless grayscale image compression. In the proposed mapping scheme, the frequency of occurrence of each symbol in the original image is computed. According to their corresponding frequencies, these symbols are sorted in descending order. Based on this order, each symbol is replaced by an 8-bit weighted fixed-length code. This replacement will generate an equivalent binary source with an increased length of successive identical symbols (0s or 1s). Different experiments using Lempel-Ziv lossless image compression algorithms have been conducted on the generated binary source. Results show that the newly proposed mapping scheme achieves some dramatic improvements in regards to compression ratios.

A DATA COMPRESSION METHOD USING ADAPTIVE BINARY ARITHMETIC CODING AND FUZZY LOGIC

  • Jou, Jer-Min;Chen, Pei-Yin
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1998.06a
    • /
    • pp.756-761
    • /
    • 1998
  • This paper describes an in-line lossless data compression method using adaptive binary arithmetic coding. To achieve better compression efficiency , we employ an adaptive fuzzy -tuning modeler, which uses fuzzy inference to deal with the problem of conditional probability estimation. The design is simple, fast and suitable for VLSI implementation because we adopt the table -look-up approach. As compared with the out-comes of other lossless coding schemes, our results are good and satisfactory for various types of source data.

  • PDF

A Low-Complexity and High-Quality Image Compression Method for Digital Cameras

  • Xie, Xiang;Li, Guolin;Wang, Zhihua
    • ETRI Journal
    • /
    • v.28 no.2
    • /
    • pp.260-263
    • /
    • 2006
  • This letter proposes a new near-lossless image compression method with only one line buffer cost for a digital camera with Bayer format image. For such format data, it can provide a low average compression rate (4.24bits/pixel) with high-image quality (larger than 46.37dB where the error of every pixel is less than two). The experimental results show that the near-lossless compression method has better performance than JPEG-LS (lossless) with ${\delta}$ = 2 for a Bayer format image.

  • PDF

MPEG-4 ALS - The Standard for Lossless Audio Coding

  • Liebchen, Tilman
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.7
    • /
    • pp.618-629
    • /
    • 2009
  • The MPEG-4 Audio Lossless Coding (ALS) standard belongs to the family MPEG-4 audio coding standards. In contrast to lossy codecs such as AAC, which merely strive to preserve the subjective audio quality, lossless coding preserves every single bit of the original audio data. The ALS core codec is based on forward-adaptive linear prediction, which combines remarkable compression with low complexity. Additional features include long-term prediction, multichannel coding, and compression of floating-point audio material. This paper describes the basic elements of the ALS codec with a focus on prediction, entropy coding, and related tools and points out the most important applications of this standardized lossless audio format.

Lossless Image Compression Using Block-Adaptive Context Tree Weighting (블록 적응적인 Context Tree Weighting을 이용한 무손실 영상 압축)

  • Oh, Eun-ju;Cho, Hyun-ji;Yoo, Hoon
    • Journal of Internet Computing and Services
    • /
    • v.21 no.4
    • /
    • pp.43-49
    • /
    • 2020
  • This paper proposes a lossless image compression method based on arithmetic coding using block-adaptive Context Tree Weighting. The CTW method predicts and compresses the input data bit by bit. Also, it can achieve a desirable coding distribution for tree sources with an unknown model and unknown parameters. This paper suggests the method to enhance the compression rate about image data, especially aerial and satellite images that require lossless compression. The value of aerial and satellite images is significant. Also, the size of their images is huger than common images. But, existed methods have difficulties to compress these data. For these reasons, this paper shows the experiment to prove a higher compression rate when using the CTW method with divided images than when using the same method with non-divided images. The experimental results indicate that the proposed method is more effective when compressing the divided images.

QuadTree-Based Lossless Image Compression and Encryption for Real-Time Processing (실시간 처리를 위한 쿼드트리 기반 무손실 영상압축 및 암호화)

  • Yoon, Jeong-Oh;Sung, Woo-Seok;Hwang, Chan-Sik
    • The KIPS Transactions:PartC
    • /
    • v.8C no.5
    • /
    • pp.525-534
    • /
    • 2001
  • Generally, compression and encryption procedures are performed independently in lossless image compression and encryption. When compression is followed by encryption, the compressed-stream should have the property of randomness because its entropy is decreased during the compression. However, when full data is compressed using image compression methods and then encrypted by encryption algorithms, real-time processing is unrealistic due to the time delay involved. In this paper, we propose to combine compression and encryption to reduce the overall processing time. It is method decomposing gray-scale image by means of quadtree compression algorithms and encrypting the structural part. Moreover, the lossless compression ratio can be increased using a transform that provides an decorrelated image and homogeneous region, and the encryption security can be improved using a reconstruction of the unencrypted quadtree data at each level. We confirmed the increased compression ratio, improved encryption security, and real-time processing by using computer simulations.

  • PDF