• Title/Summary/Keyword: multi-view video coding

Search Result 109, Processing Time 0.02 seconds

A Bit Allocation Method Based on Proportional-Integral-Derivative Algorithm for 3DTV

  • Yan, Tao;Ra, In-Ho;Liu, Deyang;Zhang, Qian
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.5
    • /
    • pp.1728-1743
    • /
    • 2021
  • Three-dimensional (3D) video scenes are complex and difficult to control, especially when scene switching occurs. In this paper, we propose two algorithms based on an incremental proportional-integral-derivative (PID) algorithm and a similarity analysis between views to improve the method of bit allocation for multi-view high efficiency video coding (MV-HEVC). Firstly, an incremental PID algorithm is introduced to control the buffer "liquid level" to reduce the negative impact on the target bit allocation of the view layer and frame layer owing to the fluctuation of the buffer "liquid level". Then, using the image similarity between views is used to establish, a bit allocation calculation model for the multi-view video main viewpoint and non-main viewpoint is established. Then, a bit allocation calculation method based on hierarchical B frames is proposed. Experimental simulation results verify that the algorithm ensures a smooth transition of image quality while increasing the coding efficiency, and the PSNR increases by 0.03 to 0.82dB while not significantly increasing the calculation complexity.

Post-processing of 3D Video Extension of H.264/AVC for a Quality Enhancement of Synthesized View Sequences

  • Bang, Gun;Hur, Namho;Lee, Seong-Whan
    • ETRI Journal
    • /
    • v.36 no.2
    • /
    • pp.242-252
    • /
    • 2014
  • Since July of 2012, the 3D video extension of H.264/AVC has been under development to support the multi-view video plus depth format. In 3D video applications such as multi-view and free-view point applications, synthesized views are generated using coded texture video and coded depth video. Such synthesized views can be distorted by quantization noise and inaccuracy of 3D wrapping positions, thus it is important to improve their quality where possible. To achieve this, the relationship among the depth video, texture video, and synthesized view is investigated herein. Based on this investigation, an edge noise suppression filtering process to preserve the edges of the depth video and a method based on a total variation approach to maximum a posteriori probability estimates for reducing the quantization noise of the coded texture video. The experiment results show that the proposed methods improve the peak signal-to-noise ratio and visual quality of a synthesized view compared to a synthesized view without post processing methods.

Bayesian-theory-based Fast CU Size and Mode Decision Algorithm for 3D-HEVC Depth Video Inter-coding

  • Chen, Fen;Liu, Sheng;Peng, Zongju;Hu, Qingqing;Jiang, Gangyi;Yu, Mei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.4
    • /
    • pp.1730-1747
    • /
    • 2018
  • Multi-view video plus depth (MVD) is a mainstream format of 3D scene representation in free viewpoint video systems. The advanced 3D extension of the high efficiency video coding (3D-HEVC) standard introduces new prediction tools to improve the coding performance of depth video. However, the depth video in 3D-HEVC is time consuming. To reduce the complexity of the depth video inter coding, we propose a fast coding unit (CU) size and mode decision algorithm. First, an off-line trained Bayesian model is built which the feature vector contains the depth levels of the corresponding spatial, temporal, and inter-component (texture-depth) neighboring largest CUs (LCUs). Then, the model is used to predict the depth level of the current LCU, and terminate the CU recursive splitting process. Finally, the CU mode search process is early terminated by making use of the mode correlation of spatial, inter-component (texture-depth), and inter-view neighboring CUs. Compared to the 3D-HEVC reference software HTM-10.0, the proposed algorithm reduces the encoding time of depth video and the total encoding time by 65.03% and 41.04% on average, respectively, with negligible quality degradation of the synthesized virtual view.

H.264 Encoding Technique of Multi-view Video expressed by Layered Depth Image (계층적 깊이 영상으로 표현된 다시점 비디오에 대한 H.264 부호화 기술)

  • Shin, Jong-Hong;Jee, Inn-Ho
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.14 no.2
    • /
    • pp.43-51
    • /
    • 2014
  • Multi-view video including depth image is necessary to develop a new compression encoding technique for storage and transmission, because of a huge amount of data. Layered depth image is an efficient representation method of multi-view video data. This method makes a data structure that is synthesis of multi-view color and depth image. This efficient method to compress new contents is suggested to use layered depth image representation and to apply for video compression encoding by using 3D warping. This paper proposed enhanced compression method using layered depth image representation and H.264/AVC video coding technology. In experimental results, we confirmed high compression performance and good quality of reconstructed image.

Multi-View Wyner-Ziv Video Coding Based on Spatio-temporal Adaptive Estimation (시공간 적응적인 예측에 기초한 다시점 위너-지브 비디오 부호화 기법)

  • Lee, Beom-yong;Kim, Jin-soo
    • The Journal of the Korea Contents Association
    • /
    • v.16 no.6
    • /
    • pp.9-18
    • /
    • 2016
  • This paper proposes a multi-view Wyner-Ziv Video coding scheme based on spatio-temporal adaptive estimation. The proposed algorithm is designed to search for a better estimated block with joint bi-directional motion estimation by introducing weights between temporal and spatial directions, and by classifying effectively the region of interest blocks, which is based on the edge detection and the synthesis, and by selecting the reference estimation block from the effective motion vector analysis. The proposed algorithm exploits the information of a single frame viewpoint and adjacent frame viewpoints, simultaneously and then generates adaptively side information in a variety of closure, and reflection regions to have a better performance. Through several simulations with multi-view video sequences, it is shown that the proposed algorithm performs visual quality improvement as well as bit-rate reduction, compared to the conventional methods.

Flexible GGOP prediction structure for multi-view video coding (다시점 동영상 부호화를 위한 가변형 다시점GOP 예측 구조)

  • Yoon, Jae-Won;Seo, Jung-Dong;Kim, Yong-Tae;Park, Chang-Seob;Sohn, Kwang-Hoon
    • Journal of Broadcast Engineering
    • /
    • v.11 no.4 s.33
    • /
    • pp.420-430
    • /
    • 2006
  • In this paper, we propose a flexible GGOP prediction structure to improve coding efficiency for multi-view video coding. In general, reference software used for MVC uses the fixed GGOP prediction structure. However, the performance of MVC depends on the base view and numbers of B-pictures between I-picture(or P-picture) and P-picture. In order to implement the flexible GGOP prediction structure, the location of base view is decided according to the global disparities among the adjacent sequences. Numbers of B-pictures between I-picture(or P-picture) and P-picture are decided by camera arrangement such as the baseline distance among the cameras. The proposed method shows better result than the reference software of MVC. The proposed prediction structure shows considerable reduction of coded bits by 7.1%.

Implementing 3DoF+ 360 Video Compression System for Immersive Media (실감형 미디어를 위한 3DoF+ 360 비디오 압축 시스템 구현)

  • Jeong, Jong-Beom;Lee, Soonbin;Jang, Dongmin;Lee, Sangsoon;Ryu, Eun-Seok
    • Journal of Broadcast Engineering
    • /
    • v.24 no.5
    • /
    • pp.743-754
    • /
    • 2019
  • System for three degrees of freedom plus (3DoF+) and 6DoF requires multi-view high resolution 360 video transmission to provide user viewport adaptive 360 video streaming. In this paper, we implement 3DoF+ 360 video compression system which removes the redundancy between multi-view videos and merges the residual into one video to provide high quality 360 video corresponding to an user's head movement efficiently. Implementations about 3D warping based redundancy removal method between 3DoF+ 360 videos and residual extraction and merger are explained in this paper. With the proposed system, 20.14% of BD-rate reduction in maximum is shown compared to traditional high-efficiency video coding (HEVC) based system.

Digital Hologram Compression Technique using Multi-view Prediction Method (다시점 예측 기법을 이용한 디지털 홀로그램의 압축 기술)

  • Choi, Hyun-Jun;Seo, Young-Ho;Bae, Jin-Woo;Yoo, Ji-Sang;Kim, Dong-Wook
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.413-414
    • /
    • 2006
  • In this paper, we propose an efficient coding method of digital hologram (or fringe pattern) using multi-view prediction technique based on video and image coding standard such as MPEG. The proposed algorithm illustrates that it has better properties for reconstruction and higher compression rate than the previous researches.

  • PDF

Auto-Covariance Analysis for Depth Map Coding

  • Liu, Lei;Zhao, Yao;Lin, Chunyu;Bai, Huihui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.9
    • /
    • pp.3146-3158
    • /
    • 2014
  • Efficient depth map coding is very crucial to the multi-view plus depth (MVD) format of 3-D video representation, as the quality of the synthesized virtual views highly depends on the accuracy of the depth map. Depth map contains smooth area within an object but distinct boundary, and these boundary areas affect the visual quality of synthesized views significantly. In this paper, we characterize the depth map by an auto-covariance analysis to show the locally anisotropic features of depth map. According to the characterization analysis, we propose an efficient depth map coding scheme, in which the directional discrete cosine transforms (DDCT) is adopted to substitute the conventional 2-D DCT to preserve the boundary information and thereby increase the quality of synthesized view. Experimental results show that the proposed scheme achieves better performance than that of conventional DCT with respect to the bitrate savings and rendering quality.