• Title/Summary/Keyword: Texture encoding

Search Result 26, Processing Time 0.027 seconds

Comparison of Compression Schemes for Real-Time 3D Texture Mapping (실시간 3차원 텍스춰 매핑을 위한 압축기법의 성능 비교)

  • Park, Gi-Ju;Im, In-Seong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.6 no.4
    • /
    • pp.35-42
    • /
    • 2000
  • 3D texture mapping generates highly natural visual effects in which objects appear carved from lumps of materials rather than laminated with thin sheets as in 2D texture mapping. Storing 3D texture images in a table for fast mapping computations, instead of evaluating procedures on the fly, however, has been considered impractical due to the extremely high memory requirement. Recently, a practical real-time 3D texture mapping technique was proposed in [11], where they attempt to resolve the potential texture memory problem by compressing 3D textures using a wavelet-based encoding method. In this paper, we consider two other encoding schemes that could also be applied to the compression-based 3D texture mapping. In particular, we extend the vector quantization and FXT1 for 3D texture compression, and compare their performance with the wavelet-based encoding scheme.

  • PDF

The Pattern Segmentation of 3D Image Information Using FCM (FCM을 이용한 3차원 영상 정보의 패턴 분할)

  • Kim Eun-Seok;Joo Ki-See
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.5
    • /
    • pp.871-876
    • /
    • 2006
  • In this thesis, to accurately measure 3D face information using the spatial encoding patterns, the new algorithm to segment the pattern images from initial face pattern image is proposed. If the obtained images is non-homogeneous texture and ambiguous boundary pattern, the pattern segmentation is very difficult. Furthermore. the non-encoded areas by accumulated error are occurred. In this thesis, the FCM(fuzzy c-means) clustering method is proposed to enhance the robust encoding and segmentation rate under non-homogeneous texture and ambiguous boundary pattern. The initial parameters for experiment such as clustering class number, maximum repetition number, and error tolerance are set with 2, 100, 0.0001 respectively. The proposed pattern segmentation method increased 8-20% segmentation rate with conventional binary segmentation methods.

Fast Intra Mode Decision Algorithm for Depth Map Coding using Texture Information in 3D-AVC (3D-AVC에서 색상 영상 정보를 이용한 깊이 영상의 빠른 화면 내 예측 모드 결정 기법)

  • Kang, Jinmi;Chung, Kidong
    • Journal of Korea Multimedia Society
    • /
    • v.18 no.2
    • /
    • pp.149-157
    • /
    • 2015
  • The 3D-AVC standard aims at improving coding efficiency by applying new techniques for utilizing intra, inter and view predictions. 3D video scenes are rendered with existing texture video and additional depth map. The depth map comes at the expense of increased computational complexity of the encoding process. For real-time applications, reducing the complexity of 3D-AVC is very important. In this paper, we present a fast intra mode decision algorithm to reduce the complexity burden in the 3D video system. The proposed algorithm uses similarity between texture video and depth map. The best intra prediction mode of the depth map is similar to that of the corresponding texture video. The early decision algorithm can be made on the intra prediction of depth map coding by using the coded intra mode of texture video. Adaptive threshold for early termination is also proposed. Experimental results show that the proposed algorithm saves the encoding time on average 29.7% without any significant loss in terms of the bit rate or PSNR value.

Low-Complexity MPEG-4 Shape Encoding towards Realtime Object-Based Applications

  • Jang, Euee-Seon
    • ETRI Journal
    • /
    • v.26 no.2
    • /
    • pp.122-135
    • /
    • 2004
  • Although frame-based MPEG-4 video services have been successfully deployed since 2000, MPEG-4 video coding is now facing great competition in becoming a dominant player in the market. Object-based coding is one of the key functionalities of MPEG-4 video coding. Real-time object-based video encoding is also important for multimedia broadcasting for the near future. Object-based video services using MPEG-4 have not yet made a successful debut due to several reasons. One of the critical problems is the coding complexity of object-based video coding over frame-based video coding. Since a video object is described with an arbitrary shape, the bitstream contains not only motion and texture data but also shape data. This has introduced additional complexity to the decoder side as well as to the encoder side. In this paper, we have analyzed the current MPEG-4 video encoding tools and proposed efficient coding technologies that reduce the complexity of the encoder. Using the proposed coding schemes, we have obtained a 56 percent reduction in shape-coding complexity over the MPEG-4 video reference software (Microsoft version, 2000 edition).

  • PDF

Texture Classification Using Rotation Invariant Local Directional Pattern (Rotation Invariant Local Directional Pattern을 이용한 텍스처 분류 방법)

  • Lee, Tae Hwan;Chae, Ok Sam
    • Convergence Security Journal
    • /
    • v.17 no.3
    • /
    • pp.21-29
    • /
    • 2017
  • Accurate encoding of local patterns is a very important factor in texture classification. However, LBP based methods w idely studied have fundamental problems that are vulnerable to noise. Recently, LDP method using edge response and dire ction information was proposed in facial expression recognition. LDP is more robust to noise than LBP and can accommod ate more information in it's pattern code, but it has drawbacks that it is sensitive to rotation transforms that are critical to texture classification. In this paper, we propose a new local pattern coding method called Rotation Invariant Local Direc tional Pattern, which combines rotation-invariant transform to LDP. To prove the texture classification performance of the proposed method in this paper, texture classification was performed on the widely used UIUC and CUReT datasets. As a result, the proposed RILDP method showed better performance than the existing methods.

ASTC Block-Size Determination Method based on PSNR Values (PSNR 값 기반의 자동화된 ASTC 블록 크기 결정 방법)

  • Nah, Jae-Ho
    • Journal of the Korea Computer Graphics Society
    • /
    • v.28 no.2
    • /
    • pp.21-28
    • /
    • 2022
  • ASTC is one of the standard texture formats supported in OpenGL ES 3.2 and Vulkan 1.0 (and later versions), and it has been increasingly used on mobile platforms (Android and iOS). ASTC's most important feature is the block size configuration, thereby providing a trade-off between compression quality and rates. With the higher number of textures, however, it is difficult to manually determine the optimal block sizes of each texture. To solve the problem, we present a new approach based on PSNR values to automatically determine the ASTC block size. A brute-force approach, which compresses a texture on all block sizes and compares the PSNR values of the compressed textures, can increase the compression time by up to 14 times. In contrast, our three-step approach minimizes the compression-time overhead. According to our experiments on a texture set including 64 various textures, our method determined the block sizes from 4×4 to 12×12 and reduced the size of compressed files by 68%.

Real-Time Panoramic Video Streaming Technique with Multiple Virtual Cameras (다중 가상 카메라의 실시간 파노라마 비디오 스트리밍 기법)

  • Ok, Sooyol;Lee, Suk-Hwan
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.4
    • /
    • pp.538-549
    • /
    • 2021
  • In this paper, we introduce a technique for 360-degree panoramic video streaming with multiple virtual cameras in real-time. The proposed technique consists of generating 360-degree panoramic video data by ORB feature point detection, texture transformation, panoramic video data compression, and RTSP-based video streaming transmission. Especially, the generating process of 360-degree panoramic video data and texture transformation are accelerated by CUDA for complex processing such as camera calibration, stitching, blending, encoding. Our experiment evaluated the frames per second (fps) of the transmitted 360-degree panoramic video. Experimental results verified that our technique takes at least 30fps at 4K output resolution, which indicates that it can both generates and transmits 360-degree panoramic video data in real time.

Image Coding by Block Based Fractal Approximation (블록단위의 프래탈 근사화를 이용한 영상코딩)

  • 정현민;김영규;윤택현;강현철;이병래;박규태
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.31B no.2
    • /
    • pp.45-55
    • /
    • 1994
  • In this paper, a block based image approximation technique using the Self Affine System(SAS) from the fractal theory is suggested. Each block of an image is divided into 4 tiles and 4 affine mapping coefficients are found for each tile. To find the affine mapping cefficients that minimize the error between the affine transformed image block and the reconstructed image block, the matrix euation is solved by setting each partial differential coefficients to aero. And to ensure the convergence of coding block. 4 uniformly partitioned affine transformation is applied. Variable block size technique is employed in order to applynatural image reconstruction property of fractal image coding. Large blocks are used for encoding smooth backgrounds to yield high compression efficiency and texture and edge blocks are divided into smaller blocks to preserve the block detail. Affine mapping coefficinets are found for each block having 16$\times$16, 8$\times$8 or 4$\times$4 size. Each block is classified as shade, texture or edge. Average gray level is transmitted for shade bolcks, and coefficients are found for texture and edge blocks. Coefficients are quantized and only 16 bytes per block are transmitted. Using the proposed algorithm, the computational load increases linearly in proportion to image size. PSNR of 31.58dB is obtained as the result using 512$\times$512, 8 bits per pixel Lena image.

  • PDF

An Algorithm for the Multi-view Image Improvement with the Resteicted Number of Images in Texture Extraction (텍스쳐 추출시 제한된 수의 참여 영상을 이용한 Multi-view 영상 개선 알고리듬)

  • 김도현;양영일
    • Journal of Korea Multimedia Society
    • /
    • v.3 no.1
    • /
    • pp.34-40
    • /
    • 2000
  • '[n this paper, we propose an efficient multi-view image coding algorithm which finds the optimal texture from a restricted number of multi-view image. The X-Y plane of the normalized object space is divided into the triangular patches. The depth of each node is determined by appling a block based disparity compensation method. Thereafter the texture of each patch is extracted by appling an affine transformation based disparity compensation method to the multi-view images. We reduced the number of images needed to determine the texture compared to traditional methods which use all the multi-view image in the texture extraction. The experimental results show that the SNR of images encoded by the proposed algorithm is better than that of images encoded by the traditional method by the approximately 0.2dB for the test sets of multi -view image called dragon, santa, city and kid. Image data recovered after encoding by the proposed method show a better visual results than after using traditional method.

  • PDF

Bayesian-theory-based Fast CU Size and Mode Decision Algorithm for 3D-HEVC Depth Video Inter-coding

  • Chen, Fen;Liu, Sheng;Peng, Zongju;Hu, Qingqing;Jiang, Gangyi;Yu, Mei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.4
    • /
    • pp.1730-1747
    • /
    • 2018
  • Multi-view video plus depth (MVD) is a mainstream format of 3D scene representation in free viewpoint video systems. The advanced 3D extension of the high efficiency video coding (3D-HEVC) standard introduces new prediction tools to improve the coding performance of depth video. However, the depth video in 3D-HEVC is time consuming. To reduce the complexity of the depth video inter coding, we propose a fast coding unit (CU) size and mode decision algorithm. First, an off-line trained Bayesian model is built which the feature vector contains the depth levels of the corresponding spatial, temporal, and inter-component (texture-depth) neighboring largest CUs (LCUs). Then, the model is used to predict the depth level of the current LCU, and terminate the CU recursive splitting process. Finally, the CU mode search process is early terminated by making use of the mode correlation of spatial, inter-component (texture-depth), and inter-view neighboring CUs. Compared to the 3D-HEVC reference software HTM-10.0, the proposed algorithm reduces the encoding time of depth video and the total encoding time by 65.03% and 41.04% on average, respectively, with negligible quality degradation of the synthesized virtual view.