• Title/Summary/Keyword: 비디오 영상 부호화

Search Result 445, Processing Time 0.027 seconds

Fast Coding Mode Decision for Temporal Scalability in H.264/AVC Scalable Extension (시간적 계층에서의 스케일러블 부호화 고속 모드 결정 방법)

  • Jeon, Byeungwoo
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.6 no.2
    • /
    • pp.71-75
    • /
    • 2013
  • Recently proliferating heterogeneous multimedia service environments should be able to deal with many different transmission speeds, image sizes, or qualities of video. However, not many existing video compression standards satisfy those necessities. To satisfy the functional requirements, the standardization of the H.264/AVC Scalable Extension (SE) technique has been recently completed. It is an extension of the H.264/AVC which can encode several image sizes and qualities at the same time as a single bitstream. To perform optimum mode decision, motion estimation is performed for all MB modes, and the RD costs are compared to identify an MB mode with the smallest RD cost. This increases computational complexity of H.264/AVC SE encoding. In this paper, we propose an early skip mode detection scheme to reduce candidate modes and suggest an algorithm of fast mode decision utilizing reference modes according to the mode history.

Scalable Video Coding using Super-Resolution based on Convolutional Neural Networks for Video Transmission over Very Narrow-Bandwidth Networks (초협대역 비디오 전송을 위한 심층 신경망 기반 초해상화를 이용한 스케일러블 비디오 코딩)

  • Kim, Dae-Eun;Ki, Sehwan;Kim, Munchurl;Jun, Ki Nam;Baek, Seung Ho;Kim, Dong Hyun;Choi, Jeung Won
    • Journal of Broadcast Engineering
    • /
    • v.24 no.1
    • /
    • pp.132-141
    • /
    • 2019
  • The necessity of transmitting video data over a narrow-bandwidth exists steadily despite that video service over broadband is common. In this paper, we propose a scalable video coding framework for low-resolution video transmission over a very narrow-bandwidth network by super-resolution of decoded frames of a base layer using a convolutional neural network based super resolution technique to improve the coding efficiency by using it as a prediction for the enhancement layer. In contrast to the conventional scalable high efficiency video coding (SHVC) standard, in which upscaling is performed with a fixed filter, we propose a scalable video coding framework that replaces the existing fixed up-scaling filter by using the trained convolutional neural network for super-resolution. For this, we proposed a neural network structure with skip connection and residual learning technique and trained it according to the application scenario of the video coding framework. For the application scenario where a video whose resolution is $352{\times}288$ and frame rate is 8fps is encoded at 110kbps, the quality of the proposed scalable video coding framework is higher than that of the SHVC framework.

An Adaptive Rate Control Using Piecewise Linear Approximation Model (부분 선형 근사 모델을 이용한 적응적 비트율 제어)

  • 조창형;정제창;최병욱
    • Journal of Broadcast Engineering
    • /
    • v.2 no.2
    • /
    • pp.194-205
    • /
    • 1997
  • In video compression standards such as MPEG and H.263. rate control is one of the key components for good coding performance. This paper presents a simple adaptive rate control scheme using a piecewise linear approximation model. While conventional buffer control approach is performed by adjusting the quantization parameter linearly according to the buffer fullness. the proposed approach uses a piecewise linear approximation model derived from logarithmic relation between the quantization parameter and bitrate in data compression. In addition. a forward analyzer performed in the spatial domain is used to improve image quality. Simulation results demonstrate that the proposed method provides better performance than the conventional one and reduces the fluctuation of the PSNR per frame while maintaining the quality of the reconstructed frames at a relatively stable level.

  • PDF

H.264/SVC Spatial Scalability Coding based Terrestrial Multi-channel Hybrid HD Broadcasting Service Framework and Performance Analysis on H.264/SVC (H.264/SVC 공간 계위 부호화 기반 지상파 다채널 하이브리드 고화질 방송 서비스 프레임워크 및 H.264/SVC 부호화 성능 평가)

  • Kim, Dae-Eun;Lee, Bum-Shik;Kim, Mun-Churl;Kim, Byung-Sun;Hahm, Sang-Jin;Lee, Keun-Sik
    • Journal of Broadcast Engineering
    • /
    • v.17 no.4
    • /
    • pp.640-658
    • /
    • 2012
  • One of the existing terrestrial multi-channel DTV service frameworks, called KoreaView, provides four programs, composed of MPEG-2 based one HD video and H.264/AVC based three SD videos within one single 6MHz frequency bandwidth. However the additional 3 SD videos can not provide enough quality due to its reduced spatial resolution and low target bitrates. In this paper, we propose a framework, which is called a terrestrial multi-channel high quality hybrid DTV service, to overcome such a weakness of KoreaView services. In the proposed framework, the three additional SD videos are encoded based on an H.264/SVC Spatial Base layer, which is compliant with H.264/AVC, and are delivered via broadcasting networks. On the other hand, and the corresponding three additional HD videos are encoded based on an H.264/SVC Spatial Enhancement layer, which are transmitted over broadband networks such as Internet, thus allowing the three additional videos for users with better quality of experience. In order to verify the effectiveness of the proposed framework, various experimental results are provided for real video contents being used for DTV services. First, the experimental results show that, when the SD sequences are encoded by the H.264/SVC Spatial Base layer at a target bitrate of 1.5Mbps, the resulting PSNR values are ranged from 34.5dB to 42.9dB, which is a sufficient level of service quality. Also it is noted that 690kbps-8,200kbps are needed for the HD test sequences when they are encoded in the H.264/SVC Spatial Enhancement layer at similar PSNR values for the same HD sequences encoded by MPEG-2 at a target bitrate of 12 Mbps.

Residual Signal Transform for Digital Cinema Sequences Lossless Coding (디지털 시네마 영상 무손실 부호화를 위한 전차신호 변환)

  • Han, Ki-Hun;Lee, Yung-Lyul
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2006.11a
    • /
    • pp.73-76
    • /
    • 2006
  • H.264는 MPEG-2, MPEG-4 Part 2, H.263 등 기존의 비디오 압축 표준들에 비해 우수한 화질과 부호화 효율을 제공하여 차세대 비디오 압축 표준으로서 널리 사용될 전망이다. 현재 H.264표준화 그룹인 JVT에서는 디지털 시네마 영상을 위한 Advanced 4:4:4 프로파일에 대한 표준화가 진행 중이다. 이 프로파일은 기존의 프로파일과 달리 화소당 8-12비트의 영상을 지원하며, YUV 영상대신 RGB 영상을 입력영상으로 사용한다. 디지털 시네마 영상은 보통 HD급 이상의 화면 크기를 가지며 초당 24Hz의 프레임율을 가진다. 이러한 영상에서는 화소간의 공간적 유사성이 매우 높아지는 경향이 있으며, 30Hz 영상에 비해 시간적 유사성이 감소하는 경향이 있다. 그 결과 H.264로 디지털 시네마 영상 압축 시, 공간 예측을 통한 Intra 매크로블록의 비율이 다른 테스트 영상들에 비해 월등히 높다는 것을 알 수 있다. 이는 디지털 시네마 영상 압축 시, 화면간 예측을 수행하는 ME/MC에 비해 공간 예측이 효율적이다는 것을 입증한다. 화면간 예측의 성능을 향상하기 위해 본 논문에서는 ME/MC 후 생성된 잔차 신호들을 간단히 변환하는 방법을 제안한다. 간단한 변환 기술이 추가되어 화면간 예측의 압축 성능이 향상됨은 물론, Inter 프레임에서 화면간 예측과 공간예측을 모두 사용하였을 때 전체적인 압축성능이 향상함을 실험을 통하여 확인 할 수 있었다.

  • PDF

HEVC Encoder Optimization using Depth Information (깊이정보를 이용한 HEVC의 인코더 고속화 방법)

  • Lee, Yoon Jin;Bae, Dong In;Park, Gwang Hoon
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.640-655
    • /
    • 2014
  • Many of today's video systems have additional depth camera to provide extra features such as 3D support. Thanks to these changes made in multimedia system, it is now much easier to obtain depth information of the video. Depth information can be used in various areas such as object classification, background area recognition, and so on. With depth information, we can achieve even higher coding efficiency compared to only using conventional method. Thus, in this paper, we propose the 2D video coding algorithm which uses depth information on top of the next generation 2D video codec HEVC. Background area can be recognized with depth information and by performing HEVC with it, coding complexity can be reduced. If current CU is background area, we propose the following three methods, 1) Earlier stop split structure of CU with PU SKIP mode, 2) Limiting split structure of CU with CU information in temporal position, 3) Limiting the range of motion searching. We implement our proposal using HEVC HM 12.0 reference software. With these methods results shows that encoding complexity is reduced more than 40% with only 0.5% BD-Bitrate loss. Especially, in case of video acquired through the Kinect developed by Microsoft Corp., encoding complexity is reduced by max 53% without a loss of quality. So, it is expected that these techniques can apply real-time online communication, mobile or handheld video service and so on.

Effects of LDPCA Frame Size for Parity Bit Estimation Methods in Fast Distributed Video Decoding Scheme (고속 분산 비디오 복호화 기법에서 패리티 비트 예측방식에 대한 LDPCA 프레임 크기 효과)

  • Kim, Man-Jae;Kim, Jin-Soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.8
    • /
    • pp.1675-1685
    • /
    • 2012
  • DVC (Distributed Video Coding) technique plays an essential role in providing low-complexity video encoder. But, in order to achieve the better rate-distortion performances, most DVC systems need feedback channel for parity bit control. This causes the DVC-based system to have high decoding latency and becomes as one of the most critical problems to overcome for a real implementation. In order to overcome this problem and to accelerate the commercialization of the DVC applications, this paper analyzes an effect of LDPCA frame size for adaptive LDPCA frame-based parity bit request estimations. First, this paper presents the LDPCA segmentation method in pixel-domain and explains the temporal-based bit request estimation method and the spatial-based bit request estimation method using the statistical characteristics between adjacent LDPCA frames. Through computer simulations, it is shown that the better performance and fast decoding is observed specially when the LDPCA frame size is 3168 in QCIF resolution.

Carriage of Multiview Video Plus Depth over MPEG-2 TS (다시점 비디오 및 깊이 영상의 MPEG-2 TS 전송기법)

  • Baek, Doo-San;Kim, Jae-Gon;Lee, Bong-Ho;Lee, Jin-Young;Yun, Kug-Jin;Cheong, Won-Sik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2011.11a
    • /
    • pp.270-272
    • /
    • 2011
  • 다시점 비디오는 사용자들에게 자유로운 시점 선택 및 넓은 시점에서의 입체감을 제공하지만, 시점의 증가에 따른 데이터 량의 증가가 불가피하다. 이에 따라 최근 MPEG 에서는 재생되는 시점 수 보다 적은 수의 다시점 비디오와 그에 대응하는 깊이 영상 및 관련 파라미터를 메타데이터 형태로 제공하는 다시점 및 깊이 영상 부호화(MVD: Multiview plus Depth)에 대한 표준화를 진행하고 있다. 본 논문은 MVD 가 방송망에서 서비스되는 시나리오를 가정하여 다시점 비디오와 그에 대응하는 깊이 영상을 MPEG-2 TS(Transport Stream)로 전송하기 위한 기법을 제시한다. 제안한 기법은 기존의 단일 영상 및 깊이 영상을 전송하기 위한 부가비디오스트림 서술자(Auxiliary Video Stream Descriptor) 및 MPEG-C Part 3 표준을 확장하여 다시점 비디오와 그에 대응하는 깊이 영상간의 매핑 정보 및 각 시점의 깊이 파라미터를 제공할 수 있다.

  • PDF