• Title/Summary/Keyword: GPU based warping

Search Result 4, Processing Time 0.021 seconds

Efficient Tiled Stereo Display System for Tangible Meeting

  • Kim, Ig-Jae;Ahn, Sang-Chul;Kim, Hyoung-Gon
    • 한국정보디스플레이학회:학술대회논문집
    • /
    • 2007.08b
    • /
    • pp.1239-1241
    • /
    • 2007
  • In this paper, we present a tiled display system for tangible meeting. We built our system as a distributed system and use GPU based warping and image blending technique for real-time processing. For efficiency, we update specific area only, where the remote user exist, in real-time and blended it with static panoramic image of remote site.

  • PDF

Tiled Stereo Display System for Immersive Telemeeting

  • Kim, Ig-Jae;Ahn, Sang-Chul;Kim, Hyoung-Gon
    • Journal of Information Display
    • /
    • v.8 no.4
    • /
    • pp.27-31
    • /
    • 2007
  • In this paper, we present an efficient tiled stereo display system for tangible meeting. For tangible meeting, it is important to provide immersive display with high resolution image to cover up the field of view and provide to the local user the same environment as that of remote site. To achieve these, a high resolution image needs to be transmitted for reconstruction of remote world, and it should be displayed using a tiled display. However, it is hard to transmit high resolution image in real time due to the limit of network bandwidth, and so we receive multiple images and reconstruct a remote world with received images in advance. Then, we update only a specific area where remote user exists by receiving low resolution image in realtime. We synthesize the transmitted image to the existing environmental map of remote world and display it as a stereo image. For this, we developed a new system which supports GPU based real time warping and blending, automatic feature extraction using machine vision technique.

View synthesis with sparse light field for 6DoF immersive video

  • Kwak, Sangwoon;Yun, Joungil;Jeong, Jun-Young;Kim, Youngwook;Ihm, Insung;Cheong, Won-Sik;Seo, Jeongil
    • ETRI Journal
    • /
    • v.44 no.1
    • /
    • pp.24-37
    • /
    • 2022
  • Virtual view synthesis, which generates novel views similar to the characteristics of actually acquired images, is an essential technical component for delivering an immersive video with realistic binocular disparity and smooth motion parallax. This is typically achieved in sequence by warping the given images to the designated viewing position, blending warped images, and filling the remaining holes. When considering 6DoF use cases with huge motion, the warping method in patch unit is more preferable than other conventional methods running in pixel unit. Regarding the prior case, the quality of synthesized image is highly relevant to the means of blending. Based on such aspect, we proposed a novel blending architecture that exploits the similarity of the directions of rays and the distribution of depth values. By further employing the proposed method, results showed that more enhanced view was synthesized compared with the well-designed synthesizers used within moving picture expert group (MPEG-I). Moreover, we explained the GPU-based implementation synthesizing and rendering views in the level of real time by considering the applicability for immersive video service.

Acceleration of Feature-Based Image Morphing Using GPU (GPU를 이용한 특징 기반 영상모핑의 가속화)

  • Kim, Eun-Ji;Yoon, Seung-Hyun;Lee, Jieun
    • Journal of the Korea Computer Graphics Society
    • /
    • v.20 no.2
    • /
    • pp.13-24
    • /
    • 2014
  • In this study, a graphics-processing-unit (GPU)-based acceleration technique is proposed for the feature-based image morphing. This technique uses the depth-buffer of the graphics hardware to calculate efficiently the shortest distance between a pixel and the control lines. The pairs of control lines between the source image and the destination image are determined by user's input, and the distance function of each control line is rendered using two rectangles and two cones. The distance between each pixel and its nearest control line is stored in the depth buffer through the graphics pipeline, and this is used to conduct the morphing operation efficiently. The pixel-unit morphing operation is parallelized using the compute unified device architecture (CUDA) to reduce the morphing time. We demonstrate the efficiency of the proposed technique using several experimental results.