• Title/Summary/Keyword: memory compression

Search Result 255, Processing Time 0.027 seconds

3D volumetric medical image coding using unbalanced tree structure (불균형 트리 구조를 이용한 3차원 의료 영상 압축)

  • Kim Young-Seop
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.7 no.4
    • /
    • pp.567-574
    • /
    • 2006
  • This paper focuses on lossy medical image compression methods for medical images that operate on three-dimensional(3-D) irreversible integer wavelet transform. We offer an application of unbalanced tree structure algorithm to medical images, using a 3-D unbalanced wavelet decomposition and a 3-D unbalanced spatial dependence tree. The wavelet decomposition is accomplished with integer wavelet filters implemented with the lifting method. We have tested our encoder on volumetric medical images using different integer filters and 16 coding unit size. The coding unit sizes of 16 slices save considerable dynamic memory(RAM) and coding delay from full sequence coding units used in previous works. If we allow the formation of trees of different lengths, then we can accomodate more transaxial scales than three. Then the encoder and decoder can then keep track of the length of the tree in which each pixel resides through the sequence of decompositions. Results show that, even with these small coding units, our algorithm with I(5,3)filter performs as well and better in lossy coding than previous coding systems using 3-D integer unbalanced wavelet transforms on volumetric medical images.

  • PDF

Sample-Adaptive Product Quantization and Design Algorithm (표본 적응 프러덕트 양자화와 설계 알고리즘)

  • 김동식;박섭형
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.12B
    • /
    • pp.2391-2400
    • /
    • 1999
  • Vector quantizer (VQ) is an efficient data compression technique for low bit rate applications. However, the major disadvantage of VQ is its encoding complexity which increases dramatically as the vector dimension and bit rate increase. Even though one can use a modified VQ to reduce the encoding complexity, it is nearly impossible to implement such a VQ at a high bit rate or for a large vector dimension because of the enormously large memory requirement for the codebook and the very large training sequence (TS) size. To overcome this difficulty, in this paper we propose a novel structurally constrained VQ for the high bit rate and the large vector dimension cases in order to obtain VQ-level performance. Furthermore, this VQ can be extended to the low bit rate applications. The proposed quantization scheme has a form of feed-forward adaptive quantizer with a short adaptation period. Hence, we call this quantization scheme sample-adaptive product quantizer (SAPQ). SAPQ can provide a 2 ~3dB improvement over the Lloyd-Max scalar quantizers.

  • PDF

An Efficient Real-time Rendering Method for Compressed Terrain Dataset with Wavelet Transform (웨이블릿 변환으로 압축된 지형 데이터의 효율적인 실시간 렌더링 기법)

  • Kim, Tae-Gwon;Lee, Eun-Seok;Shin, Byeong-Seok
    • Journal of Korea Game Society
    • /
    • v.14 no.4
    • /
    • pp.45-52
    • /
    • 2014
  • We cannot load the entire data for high-resolution terrain model to the GPU memory since its size is too big. Out-of-core approaches are commonly used to solve the problem. However, due to limited bandwidth of the secondary storage, it is difficult to render the terrain in real-time. A method that compresses the DEM data with wavelet transform on GPU, and renders the decoded data is suggested. However, it is inefficient since it has to sample the values from textures, convert them to vertices, and generate a mesh periodically. We propose a method to store the approximation coefficients of wavelet compression as vertex attributes and render the terrain by decoding the data on geometric shader. It can reduce the amount of transferring terrain texture since approximation coefficients are given as an attribute of the vertex. Also, it generate meshes without additional upload of terrain texture.

Design of Low Power H.264 Decoder Using Adaptive Pipeline (적응적 파이프라인을 적용한 저전력 H.264 복호기 설계)

  • Lee, Chan-Ho
    • Journal of the Institute of Electronics Engineers of Korea SD
    • /
    • v.47 no.9
    • /
    • pp.1-6
    • /
    • 2010
  • H.264 video coding standard is widely used due to the high compression rate and quality. H.264 decoders usually have pipeline architecture by a macroblock or a $4{\times}4$ sub-block. The period of the pipeline is usually fixed to guarantee the operation in the worst case which results in many idle cycles and the requirement of high data bandwidth and high performance processing units. We propose adaptive pipeline architecture for H.264 decoders for efficient decoding and lower the requirement of the bandwidth for the memory bus. Parameters and coefficients are delivered using hand-shaking communication through dedicated interconnections and frame pixel data are transferred using AMBA AHB network. The processing time of each block is variable depending on the characteristics of images, and the processing units start to work whenever they are ready. An H.264 decoder is designed and implemented using the proposed architecture to verify the operation using an FPGA.

qtar: Design and Implementation of an Optimized tar Command with FTL-level Remapping (qtar: 플래시 변환 계층 리매핑 기법을 이용한 최적화된 tar 명령어 구현)

  • Ryoo, Jeongseok;Hahn, Sangwook Shane;Kim, Jihong
    • Journal of KIISE
    • /
    • v.45 no.1
    • /
    • pp.9-14
    • /
    • 2018
  • Tar is a Linux command that combines several files into a single file. Combining multiple small files into large files increases the compression efficiency and data transfer speed. However, tar has a problem in that smaller target files, result in a lower performance. In this paper, we show that this performance degradation occurs when tar reads the data from the target files and propose qtar (quick tar) to solve this problem via flash-level remapping. When the size of an I/O request is less than 1 MB, the I/O performance decreases proportionally to the decrease in size of the I/O request. Since tar reads the data of files one by one, a smaller file size results in a lower performance. Therefore, the remapping technique is implemented in qtar to read data from the target files at the maximum I/O size regardless of the size of each file. Our evaluations show that the execution time with qtar is reduced by up to 3.4 times compared to that with tar.

Key Frame Extraction and Region Segmentation-based Video Retrieval in Compressed Domain (압축영역에서의 대표프레임 추출 및 영역분할기반 비디오 검색 기법)

  • 강응관;김성주;송호근;최종수
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.9B
    • /
    • pp.1713-1720
    • /
    • 1999
  • This paper presents a new key frame extraction technique, for scene change detection, using the proposed AHIM (Accumulative Histogram Intersection Measure) from the DC image constructed by DCT DC coefficients in the compressed video sequence that is video compression standard such as MPEG. For fast content-based browsing and video retrieval in a video database, we also provide a novel coarse-to-fine video indexing scheme. In the extracted key frame, we perform the region segmentation as a preprocessing. First, the segmented image is projected with the horizontal direction, then we transform the result into a histogram, which is saved as a database index. In the second step, we calculate the moments and change them into a distance value. From the simulation results, the proposed method clearly shows the validity and superiority in respect of computation time and memory space, and that in conjunction with other techniques for indexing, such as color, can provide a powerful framework for image indexing and retrieval.

  • PDF

A Study on the Characteristic of Interaction Model for Implementation of Richmedia Contents (리치미디어 컨텐츠 구현에 있어 상호작용 모델)

  • 김민수
    • Archives of design research
    • /
    • v.17 no.1
    • /
    • pp.201-210
    • /
    • 2004
  • The web as a sign synthesis text has become a kernel for incorporeal knowledge as well as a communication model through the ubiquitous environment all over the world. The evaluation of the communication model, which is essential for the information structure, acts as an important basis on determining the quality of the web contents. In this study, the development of the progress of the communication of semantic meaning in the construction of the information structure was analyzed in views of the form, the function, and the emotional effect of the rich media contents of the web. The transformation process from the initial access elements through the final selection elements was suggested as the communication model and the effects of the function of the information in the web on the process was assessed by the engineering and linguistic models of Shannon, Weaver, and Roman Jakobson. The results of this study showed that the environments such as the speed, the memory space, data compression technique, and data filtering have influences on the web contents expression and the evaluation of the communication model in connection with the environments is the basis in the information structure.

  • PDF

Optimization of image data for Mobile Game Gontents (모바일게임 콘텐츠 개발을 위한 이미지 데이터 최적화)

  • Lee, Hwan-joong;Kim, young-bong
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2008.05a
    • /
    • pp.38-42
    • /
    • 2008
  • In the 2000s, the PC package game market has withered, but the internet based on-line game, the wireless internet based mobile game and high-powered console game market have been main stream. Among those markets, mobile game market has rapidly increased because it has 'mobility' that overcomes the limitation of time and area. But, unlike other platforms, mobile game user have to pay the higher price for download mobile game contents through the wireless internet and mobile game developer have to overcome the limitation of storage memory capacity.of mobile phone. Chiefly, the image data consume the storage capacity of mobile game contents, this paper present a technology to optimize image data for mobile game contents through analyzing type of compression method and image formats.

  • PDF

A Hiding Original App Method by Altering Android DexClassLoader Execution Flow (안드로이드 덱스클래스로더 실행흐름 변경을 통한 원본 앱 노출 방지 기법)

  • Jo, Honglae;Choi, Joontae;Ha, Dongsoo;Shin, Jinseop;Oh, Heekuck
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.27 no.6
    • /
    • pp.1271-1280
    • /
    • 2017
  • The android packer service using runtime execution compression technology switches to the original application using DexClassLoader. However the API interface of the DexClassLoader receives the path of the loaded DEX(Dalvik EXcutable) and the path of the compiled file. So there is a problem that the original file is exposed to the file system. Therefore, it is not safe to use the API for the packer service. In this paper, we solve this problem by changing the compile and load flow of the DexClassLoader API. Due to this changed execution flow, the complied file can be encrypted and stored in the file system or only in the memory and it can be decrypted or substituted at the time of subsequent loading to enable the original application conversion. we expected that the stability of the packer will increase beacause the proposed method does not expose the original file to the file system.

Design of A Deblocking Filter Based on Macroblock Overlap Scheme for H.264/AVC (H.264/AVC용 매크로블록 겹침 기법에 기반한 디블록킹 필터의 설계)

  • Kim, Won-Sam;Sonh, Seung-Il
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.12 no.4
    • /
    • pp.699-706
    • /
    • 2008
  • H.264/AVC is a new international standard for the compression of video images, in which a deblocking filter has been adopted to remoye blocking artifacts. This paper proposes an efficient architecture of deblocking filter in H.264/AVC. By making good use of data dependence between neighboring $4{\times}4$ blocks, the memory sire is reduced and the throughput of the deblocking filter processing is increased. The designed deblocking filter further enhances the parallelism by simultaneously executing horizontal and vertical filtering within a macroblock in pipeline method and adopting overlap between macroblocks. The implementation result shows that the proposed architecture enhances the performance of deblocking filter processing from 1.75 to 4.23 times than that of the conventional deblocking filter. Hence the Proposed architecture of deblocking filter is able to perform real-time deblocking in high-resolution($2048{\times}1024$) video applications.