• Title/Summary/Keyword: 2-D Coding

Search Result 564, Processing Time 0.026 seconds

Dependent Quantization for Scalable Video Coding

  • Pranantha, Danu;Kim, Mun-Churl;Hahm, Sang-Jin;Lee, Keun-Sik;Park, Keun-Soo
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2006.11a
    • /
    • pp.127-132
    • /
    • 2006
  • Quantization in video coding plays an important role in controlling the bit-rate of compressed video bit-streams. It has been used as an important control means to adjust the amount of bit-streams to at]owed bandwidth of delivery networks and storage. Due to the dependent nature of video coding, dependent quantization has been proposed and applied for MPEG-2 video coding to better maintain the quality of reconstructed frame for given constraints of target bit-rate. Since Scalable Video Coding (SVC) being currently standardized exhibits highly dependent coding nature not only between frames but also lower and higher scalability layers where the dependent quantization can be effectively applied, in this paper, we propose a dependent quantization scheme for SVC and compare its performance in visual qualities and bit-rates with the current JSVM reference software for SVC. The proposed technique exploits the frame dependences within each GOP of SVC scalability layers to formulate dependent quantization. We utilize Lagrange optimization, which is widely accepted in R-D (rate-distortion) based optimization, and construct trellis graph to find the optimal cost path in the trellis by minimizing the R-D cost. The optimal cost path in the trellis graph is the optimal set of quantization parameters (QP) for frames within a GOP. In order to reduce the complexity, we employ pruning procedure using monotonicity property in the trellis optimization and cut the frame dependency into one GOP to decrease dependency depth. The optimal Lagrange multiplier that is used for SVC is equal to H.264/AVC which is also used in the mode prediction of the JSVM reference software. The experimental result shows that the dependent quantization outperforms the current JSVM reference software encoder which actually takes a linear increasing QP in temporal scalability layers. The superiority of the dependent quantization is achieved up to 1.25 dB increment in PSNR values and 20% bits saving for the enhancement layer of SVC.

  • PDF

A study on motion prediction and subband coding of moving pictuers using GRNN (GRNN을 이용한 동영상 움직임 예측 및 대역분할 부호화에 관한 연구)

  • Han, Young-Oh
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.5 no.3
    • /
    • pp.256-261
    • /
    • 2010
  • In this paper, a new nonlinear predictor using general regression neural network(GRNN) is proposed for the subband coding of moving pictures. The performance of a proposed nonlinear predictor is compared with BMA(Block Match Algorithm), the most conventional motion estimation technique. As a result, the nonlinear predictor using GRNN can predict well more 2-3dB than BMA. Specially, because of having a clustering process and smoothing noise signals, this predictor well preserves edges in frames after predicting the subband signal. This result is important with respect of human visual system and is excellent performance for the subband coding of moving pictures.

HEVC Encoder Optimization using Depth Information (깊이정보를 이용한 HEVC의 인코더 고속화 방법)

  • Lee, Yoon Jin;Bae, Dong In;Park, Gwang Hoon
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.640-655
    • /
    • 2014
  • Many of today's video systems have additional depth camera to provide extra features such as 3D support. Thanks to these changes made in multimedia system, it is now much easier to obtain depth information of the video. Depth information can be used in various areas such as object classification, background area recognition, and so on. With depth information, we can achieve even higher coding efficiency compared to only using conventional method. Thus, in this paper, we propose the 2D video coding algorithm which uses depth information on top of the next generation 2D video codec HEVC. Background area can be recognized with depth information and by performing HEVC with it, coding complexity can be reduced. If current CU is background area, we propose the following three methods, 1) Earlier stop split structure of CU with PU SKIP mode, 2) Limiting split structure of CU with CU information in temporal position, 3) Limiting the range of motion searching. We implement our proposal using HEVC HM 12.0 reference software. With these methods results shows that encoding complexity is reduced more than 40% with only 0.5% BD-Bitrate loss. Especially, in case of video acquired through the Kinect developed by Microsoft Corp., encoding complexity is reduced by max 53% without a loss of quality. So, it is expected that these techniques can apply real-time online communication, mobile or handheld video service and so on.

3D 비디오 부호화 표준 기술

  • Park, Si-Nae;Sim, Dong-Gyu
    • The Magazine of the IEIE
    • /
    • v.37 no.9
    • /
    • pp.33-41
    • /
    • 2010
  • 최근 디스플레이 기술의 비약적인 발전과 함께 3D 영화의 흥행으로 인해 국내 뿐 아니라 전 세계적으로 3DTV에 대한 관심이 높아지고 있다. 3DTV는 사람의 눈 사이의 간격 때문에 두 눈에 맺히는 상이 달라지는 양안시차의 원리를 이용하는 기술로, 두 눈에 맺힐 두 영상을 각각 획득하고, 이를 사람의 두 눈에 각각 보여지도록 하는 방식으로 3차원 입체 비디오를 실현하게 된다. 이를 위한 3D 비디오의 부호화 표준 기술로는 기존에 MPEG-2 stereo 및 MPEG-C part 2가 ISO/IEC MPEG을 통하여 제정된바 있으며, 최근에는 ITU-T의 VCEG과 ISO/IEC MPEG이 비디오 부호화 표준을 위하여 Joint Video Coding (JVT)를 구성하여, Multi-view Video Coding (MVC)를 제정하였다. 그리고 현재 진행되는 3D비디오 관련 표준화로는 MPEG에서 Free view-point TV (FTV)등의 응용을 위한 3DV라는 이름으로 차세대 비디오 표준을 준비하고 있다. 본 고에서는 기존에 MPEG을 통해 진행된 3DTV 관련 표준화 기술을 알아보고, 현재 진행되고 있는 3DTV 부호화 표준화 동향을 살펴본다.

  • PDF

Adaptive Quantization for Transform Domain Wyner-Ziv Residual Coding of Video (변환 영역 Wyner-Ziv 잔차 신호 부호화를 위한 적응적 양자화)

  • Cho, Hyon-Myong;Shim, Hiuk-Jae;Jeon, Byeung-Woo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.4
    • /
    • pp.98-106
    • /
    • 2011
  • Since prediction processes such as motion estimation motion compensation are not at the WZ video encoder but at its decoder, WZ video compression cannot have better performance than that of conventional video encoder. In order to implement the prediction process with low complexity at the encoder, WZ residual coding was proposed. Instead of original WZ frames, WZ residual coding encodes the residual signal between key frames and WZ frames. Although the proposed WZ residual coding has good performance in pixel domain, it does not have any improvements in transform domain compared to transform domain WZ coding. The WZ residual coding in transform domain is difficult to have better performance, because pre-defined quantization matrices in WZ coding are not compatible with WZ residual coding. In this paper, we propose a new quantization method modifying quantization matrix and quantization step size adaptively for transform domain WZ residual coding. Experimental result shows 22% gain in BDBR and 1.2dB gain in BDPSNR.

A Frame-based Coding Mode Decision for Temporally Active Video Sequence in Distributed Video Coding (분산비디오부호화에서 동적비디오에 적합한 프레임별 모드 결정)

  • Hoangvan, Xiem;Park, Jong-Bin;Shim, Hiuk-Jae;Jeon, Byeung-Woo
    • Journal of Broadcast Engineering
    • /
    • v.16 no.3
    • /
    • pp.510-519
    • /
    • 2011
  • Intra mode decision is a useful coding tool in Distributed Video Coding (DVC) for improving DVC coding efficiency for video sequences having fast motion. A major limitation associated with the existing intra mode decision methods, however, is that its efficiency highly depends on user-specified thresholds or modeling parameters. This paper proposes an entropy-based method to address this problem. The probabilities of intra and Wyner?Ziv (WZ) modes are determined firstly by examining correlation of pixels in spatial and temporal directions. Based on these probabilities, entropy of the intra and the WZ modes are computed. A comparison based on the entropy values decides a coding mode between intra coding and WZ coding without relying on any user-specified thresholds or modeling parameters. Experimental results show its superior rate-distortion performance of improvements of PSNR up to 2 dB against a conventional Wyner?Ziv coding without intra mode decision. Furthermore, since the proposed method does not require any thresholds or modeling parameters from users, it is very attractive for real life applications.

Quasi-Complementary Turbo Codes (QCTC) for cdma2000 1xEV-DV

  • Kim, Min-Goo;Ha, Sang-Hyuk;Kim, Yong-Serk
    • Proceedings of the IEEK Conference
    • /
    • 2003.07a
    • /
    • pp.97-100
    • /
    • 2003
  • The quasi-complementary turbo codes (QCTC) proposed by Kim [1] is used for a fast hybrid ARQ scheme with incremental redundancy and adaptive modulation coding in the cdma2000 1xEV-DV [2]. The QCTC provides various code rates with good performance, a very simple encoder structure, and an inherent channel interleaving. It is shown that the QCTC is a unified scheme of channel coding and channel interleaving. In this paper, we introduce the properties of QCTC and various hybrid ARQ-QCTC schemes for the system.

  • PDF

Multi-view Video Coding using View Interpolation (영상 보간을 이용한 다시점 비디오 부호화 방법)

  • Lee, Cheon;Oh, Kwan-Jung;Ho, Yo-Sung
    • Journal of Broadcast Engineering
    • /
    • v.12 no.2
    • /
    • pp.128-136
    • /
    • 2007
  • Since the multi-view video is a set of video sequences captured by multiple array cameras for the same three-dimensional scene, it can provide multiple viewpoint images using geometrical manipulation and intermediate view generation. Although multi-view video allows us to experience more realistic feeling with a wide range of images, the amount of data to be processed increases in proportion to the number of cameras. Therefore, we need to develop efficient coding methods. One of the possible approaches to multi-view video coding is to generate an intermediate image using view interpolation method and to use the interpolated image as an additional reference frame. The previous view interpolation method for multi-view video coding employs fixed size block matching over the pre-determined disparity search range. However, if the disparity search range is not proper, disparity error may occur. In this paper, we propose an efficient view interpolation method using initial disparity estimation, variable block-based estimation, and pixel-level estimation using adjusted search ranges. In addition, we propose a multi-view video coding method based on H.264/AVC to exploit the intermediate image. Intermediate images have been improved about $1{\sim}4dB$ using the proposed method compared to the previous view interpolation method, and the coding efficiency have been improved about 0.5 dB compared to the reference model.

On Using the Human Visual System Model for Subband Coding (시각 시스템 모델을 이용한 Subband 코딩)

  • 박용철;김근숙;차일환;윤대희
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.27 no.6
    • /
    • pp.937-943
    • /
    • 1990
  • In this paper, a subband coding scheme using the human visual system(HVS) model for encoding monochrome images is proposed to produce perceptually higher quality images compared with the regular subband coding scheme. The proposed approach first transforms the intensity image to the density image by a point nonlinear transformation. A frequency band dexomposition of the density image is carried out by means of 2-D seaprable quadrature mirror filters, which split the density image spectrum into 16 equall rate subbands. Bits are allocated among the subbands to minimize the weighted mean squar error (WMSE) for differential pulse code modulation(DPCM) coding of the subbands. The weight for each subband is calculated from the modulation transfer function (MTF) of the HVS model at corresponding frequencies. The performances of the proposed approach are evaluated for 256 * 256 monochrome images at the bit rates of 0.5, 0.75 and 1.0 bita per pixel. Computer simulation results indicate that using the HVS model yields more pleasing reconstructed images than regular subband coding approach which does not use HVS model.

  • PDF

Scalable Multi-view Video Coding based on HEVC

  • Lim, Woong;Nam, Junghak;Sim, Donggyu
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.4 no.6
    • /
    • pp.434-442
    • /
    • 2015
  • In this paper, we propose an integrated spatial and view scalable video codec based on high efficiency video coding (HEVC). The proposed video codec is developed based on similarity and uniqueness between the scalable extension and 3D multi-view extension of HEVC. To improve compression efficiency using the proposed scalable multi-view video codec, inter-layer and inter-view predictions are jointly employed by using high-level syntaxes that are defined to identify view and layer information. For the inter-view and inter-layer predictions, a decoded picture buffer (DPB) management algorithm is also proposed. The inter-view and inter-layer motion predictions are integrated into a consolidated prediction by harmonizing with the temporal motion prediction of HEVC. We found that the proposed scalable multi-view codec achieves bitrate reduction of 36.1%, 31.6% and 15.8% on the top of ${\times}2$, ${\times}1.5$ parallel scalable codec and parallel multi-view codec, respectively.