• Title/Summary/Keyword: Adaptive video

Search Result 807, Processing Time 0.025 seconds

Edge-Preserving Algorithm for Block Artifact Reduction and Its Pipelined Architecture

  • Vinh, Truong Quang;Kim, Young-Chul
    • ETRI Journal
    • /
    • v.32 no.3
    • /
    • pp.380-389
    • /
    • 2010
  • This paper presents a new edge-protection algorithm and its very large scale integration (VLSI) architecture for block artifact reduction. Unlike previous approaches using block classification, our algorithm utilizes pixel classification to categorize each pixel into one of two classes, namely smooth region and edge region, which are described by the edge-protection maps. Based on these maps, a two-step adaptive filter which includes offset filtering and edge-preserving filtering is used to remove block artifacts. A pipelined VLSI architecture of the proposed deblocking algorithm for HD video processing is also presented in this paper. A memory-reduced architecture for a block buffer is used to optimize memory usage. The architecture of the proposed deblocking filter is verified on FPGA Cyclone II and implemented using the ANAM 0.25 ${\mu}m$ CMOS cell library. Our experimental results show that our proposed algorithm effectively reduces block artifacts while preserving the details. The PSNR performance of our algorithm using pixel classification is better than that of previous algorithms using block classification.

An Adaptive Control for the Propagation Errors Incurred by DCT Coefficient-Dropping Transcoder

  • Kim, Jin-Soo;Kim, Jae-Gon;Seo, Kwang-Deok;Yun, Mong-Han
    • ETRI Journal
    • /
    • v.29 no.5
    • /
    • pp.559-568
    • /
    • 2007
  • This paper presents a new distortion control scheme with a simple estimation model for the propagation errors incurred by dropping some parts of the bitstream in a frame dropping-coefficient dropping (FD-CD) transcoder. The primary goal of this paper is to facilitate bit-rate conversions and rate-distortion controls in the compressed domain without introducing a full decoding and reencoding system in the pixel domain. First, the error propagation behavior over several frame sequences due to coefficient dropping is investigated on the basis of statistical and empirical properties. Then, such properties are used to develop a simple estimation model for the CD distortion accounting for the characteristics of the underlying coded-frame. Finally, the proposed estimation model allows us to determine the amount of coefficient dropping and to effectively allocate rate-distortions into coded-frames. Experimental results show that the proposed estimation model accurately describes the characteristics of propagation errors adaptively in the compressed domain and can be easily applied to distortion control over different kinds of video sequences.

  • PDF

Night-time Vehicle Detection Based On Multi-class SVM (다중-클래스 SVM 기반 야간 차량 검출)

  • Lim, Hyojin;Lee, Heeyong;Park, Ju H.;Jung, Ho-Youl
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.10 no.5
    • /
    • pp.325-333
    • /
    • 2015
  • Vision based night-time vehicle detection has been an emerging research field in various advanced driver assistance systems(ADAS) and automotive vehicle as well as automatic head-lamp control. In this paper, we propose night-time vehicle detection method based on multi-class support vector machine(SVM) that consists of thresholding, labeling, feature extraction, and multi-class SVM. Vehicle light candidate blobs are extracted by local mean based thresholding following by labeling process. Seven geometric and stochastic features are extracted from each candidate through the feature extraction step. Each candidate blob is classified into vehicle light or not by multi-class SVM. Four different multi-class SVM including one-against-all(OAA), one-against-one(OAO), top-down tree structured and bottom-up tree structured SVM classifiers are implemented and evaluated in terms of vehicle detection performances. Through the simulations tested on road video sequences, we prove that top-down tree structured and bottom-up tree structured SVM have relatively better performances than the others.

Performance Analysis and improvement of Extension-interpolation (EI)/2D-DCT for Coding irregular Shaped object (불규칙 모양 물제의 부호화를 위한 확장-보간/2D-DCT의 성능 분석 및 개성 방안)

  • 조순제;강현수;윤병주;김성대;구본호
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.25 no.3B
    • /
    • pp.541-548
    • /
    • 2000
  • In the MPEG-4 standardization phase, many methods for coding the irregular shaped VOP (video object Plane) have been researched. Texture coding is one of interesting research items in the MPEG-4. There are the Low pass extrapolation (LPE) padding, the shape adaptive DCT (SA-DCT), and the Extension-Interpolation (EI)/2D-DCT proposed in [1] as texture coding methods. the EI/2D-DCT is the method extending and interpolating luminance values from and Arbitrarily Shaped (AS) image segment into an 8 x 8 block and transforming the extended and interpolated luminance values by the 8x8 DCT. although the EI/2D-DCT and the SA-DCT work well in coding the As image segments. they are degraded since they use one-dimensional (1-D) methods such as the 1D-EI and the 1D-DCT in the two-dimensional (2-D) space. in this paper, we analyze the performance of the EI/2D-DCTand propose a new non-symmetric sig-sag scanning method, which non-symmetrically scans the quantized coefficients in the DCT domain to improve the EI/2D-DCT.

  • PDF

Contents Adaptive MCTF Using JND (JND를 이용한 적응적 MCTF)

  • Heo, Jae-Seong;Ryu, Chul
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.1C
    • /
    • pp.48-55
    • /
    • 2009
  • In scalable video coding, MCTF plays an important role for time-scalability and SNR-scalability. But there is image quality decreasing as MCTF level is increased because time interval of each frame is extended so that is hard to find suitable motion vector. In this paper, we propose an algorithm to prevent image quality from decreasing with unsuitable motion vector during MCTF update process using JND. We adapt JND to find errors within blocks of image and set a threshold which is used to add high frequency components during update process. We can overcome time-gap between frames and achieve better image quality through the proposed algorithm.

Determination of Leaf Color and Health State of Lettuce using Machine Vision (기계시각을 이용한 상추의 엽색 및 건강상태 판정)

  • Lee, J.W.
    • Journal of Biosystems Engineering
    • /
    • v.32 no.4
    • /
    • pp.256-262
    • /
    • 2007
  • Image processing systems have been used to measure the plant parameters such as size, shape and structure of plants. There are yet some limited applications for evaluating plant colors due to illumination conditions. This study was focused to present adaptive methods to analyze plant leaf color regardless of illumination conditions. Color patches attached on the calibration bars were selected to represent leaf colors of lettuces and to test a possibility of health monitoring of lettuces. Repeatability of assigning leaf colors to color patches was investigated by two-tailed t-test for paired comparison. It resulted that there were no differences of assignment histogram between two images of one lettuce that were acquired at different light conditions. It supported that use of the calibration bars proposed for leaf color analysis provided color constancy, which was one of the most important issues in a video color analysis. A health discrimination equation was developed to classify lettuces into one of two classes, SOUND group and POOR group, using the machine vision. The classification accuracy of the developed health discrimination equation was 80.8%, compared to farmers' decision. This study could provide a feasible method to develop a standard color chart for evaluating leaf colors of plants and plant health monitoring system using the machine vision.

Adaptive Background Modeling Considering Stationary Object and Object Detection Technique based on Multiple Gaussian Distribution

  • Jeong, Jongmyeon;Choi, Jiyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.11
    • /
    • pp.51-57
    • /
    • 2018
  • In this paper, we studied about the extraction of the parameter and implementation of speechreading system to recognize the Korean 8 vowel. Face features are detected by amplifying, reducing the image value and making a comparison between the image value which is represented for various value in various color space. The eyes position, the nose position, the inner boundary of lip, the outer boundary of upper lip and the outer line of the tooth is found to the feature and using the analysis the area of inner lip, the hight and width of inner lip, the outer line length of the tooth rate about a inner mouth area and the distance between the nose and outer boundary of upper lip are used for the parameter. 2400 data are gathered and analyzed. Based on this analysis, the neural net is constructed and the recognition experiments are performed. In the experiment, 5 normal persons were sampled. The observational error between samples was corrected using normalization method. The experiment show very encouraging result about the usefulness of the parameter.

Application-Adaptive Performance Improvement in Mobile Systems by Using Persistent Memory

  • Bahn, Hyokyung
    • International journal of advanced smart convergence
    • /
    • v.8 no.1
    • /
    • pp.9-17
    • /
    • 2019
  • In this article, we present a performance enhancement scheme for mobile applications by adopting persistent memory. The proposed scheme supports the deadline guarantee of real-time applications like a video player, and also provides reasonable performances for non-real-time applications. To do so, we analyze the program execution path of mobile software platforms and find two sources of unpredictable time delays that make the deadline-guarantee of real-time applications difficult. The first is the irregular activation of garbage collection in flash storage and the second is the blocking and time-slice based scheduling used in mobile platforms. We resolve these two issues by adopting high performance persistent memory as the storage of real-time applications. By maintaining real-time applications and their data in persistent memory, I/O latency can become predictable because persistent memory does not need garbage collection. Also, we present a new scheduler that exclusively allocates a processor core to a real-time application. Although processor cycles can be wasted while a real-time application performs I/O, we depict that the processor utilization is not degraded significantly due to the acceleration of I/O by adopting persistent memory. Simulation experiments show that the proposed scheme improves the deadline misses of real-time applications by 90% in comparison with the legacy I/O scheme used in mobile systems.

CNN-based In-loop Filter on TU Block (TU 블록 크기에 따른 CNN기반 인루프필터)

  • Kim, Yang-Woo;Jeong, Seyoon;Cho, Seunghyun;Lee, Yung-Lyul
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2018.11a
    • /
    • pp.15-17
    • /
    • 2018
  • VVC(Versatile Video Coding)는 입력된 영상을 CTU(Coding Tree Unit) 단위로 분할하여 코딩하며, 이를 다시 QTBTT(Quadtree plus binary tree and triple tree)로 분할하고, TU(Transform Unit)도 이와 같은 단위로 분할된다. 따라서 TU의 크기는 $4{\times}4$, $4{\times}8$, $4{\times}16$, $4{\times}32$, $8{\times}4$, $16{\times}4$, $32{\times}4$, $8{\times}8$, $8{\times}16$, $8{\times}32$, $16{\times}8$, $32{\times}8$, $16{\times}16$, $16{\times}32$, $32{\times}16$, $32{\times}32$, $64{\times}64$의 17가지 종류가 있다. 기존의 VVC 참조 Software인 VTM에서는 디블록킹필터와 SAO(Sample Adaptive Offset)로 이루어진 인루프필터를 이용하여 에러를 복원하는데, 본 논문은 TU 크기에 따라서 원본블록과 복원블록의 차이(에러)가 통계적으로 다름을 이용하여 서로 다른 CNN(Convolution Neural Network)을 구축하고 에러를 복원하는 방법으로 VTM의 인루프 필터를 대체한다. 복원영상의 에러를 감소시키기 위하여 TU 블록크기에 따라 DenseNet의 Dense Block기반 CNN을 구성하고, Hyper Parameter와 복잡도의 감소를 위해 네트워크 간에 일부 가중치를 공유하는 모양의 Network를 구성하였다.

  • PDF

Adaptive Chroma Block Partitioning Method using Comparison of Similarity between Channels (채널 간 유사도 비교를 이용한 적응형 색차 블록 분할 방법)

  • Baek, A Ram;Choi, Sanggyu;Choi, Haechul
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2018.06a
    • /
    • pp.260-261
    • /
    • 2018
  • MPEG과 VCEG은 차세대 비디오 부호화 표준 기술 개발를 위한 JVET(Joint Video Exploration Team)을 구성하여 현재 비디오 표준화인 HEVC 대비 높은 부호화 효율을 목표로 연구를 진행하며 CfP(Call for Proposal) 단계를 진행 중이다. JVET의 공통 플랫폼인 JEM(Joint Exploration Test Model)은 HEVC의 quad-tree 기반 블록 분할 구조를 대신하여 더 많은 유연성을 제공하는 QTBT(Quad-tree plus binary-tree)가 적용되었다. QTBT는 화면 내 부호화 효율을 높이기 위한 하나의 방법으로 휘도와 색차 신호에 대해 분할된 블록 구조를 지원한다. 이러한 방법은 채널 간 블록 분할 모양이 동일하거나 비슷한 경우에 중복되는 블록 분할 신호가 발생할 수 있는 단점이 있다. 따라서 본 논문에서는 화면 내 부호화에서 채널 간 유사도 비교를 이용하여 적응형 색차 블록 방법을 제안한다. 제안한 방법의 실험 결과로 JEM 6.0과 비교하여 CfE(Call for Evidence) 영상에서 평균 0.28%의 Y BD-rate 감소와 함께 평균 124.5%의 부호화 복잡도 증가를 확인하였다.

  • PDF