• Title/Summary/Keyword: Image Rendering

Search Result 543, Processing Time 0.029 seconds

A Study on the Effective Image Sequence Format in 3D Animation Production (3D 애니메이션 제작에 있어서 효율적인 Image Sequence format에 관한 연구)

  • Kim Ho
    • Proceedings of the Korea Contents Association Conference
    • /
    • 2005.11a
    • /
    • pp.131-136
    • /
    • 2005
  • In 3D animation rendering process, Although we can render the output as a movie file format, most productions use image sequences in their rendering pipelines. This Image Sequence rendering process is extremely important step in final compositing in movie industries. Although there are various type of making image rendering processes, TGA format is one of most widely used bitmap file formats using in industries. People may ask TGA format is most suitable for in any case. As we know 3D softwares have their own image formats. so we need to testify on this. In this paper, we are going to focus on Alias' 3D package software called MAYA which we will analyze of compressing image sequence, Image quality, supporting Alpha channels in compositing, and Z-depth Information. The purpose of this paper is providing to 3D Pipeline as a guideline about effective image sequence format.

  • PDF

RAY-SPACE INTERPOLATION BYWARPING DISPARITY MAPS

  • Moriy, Yuji;Yendoy, Tomohiro;Tanimotoy, Masayuki;Fujiiz, Toshiaki
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.583-587
    • /
    • 2009
  • In this paper we propose a new method of Depth-Image-Based Rendering (DIBR) for Free-viewpoint TV (FTV). In the proposed method, virtual viewpoint images are rendered with 3D warping instead of estimating the view-dependent depth since depth estimation is usually costly and it is desirable to eliminate it from the rendering process. However, 3D warping causes some problems that do not occur in the method with view-dependent depth estimation; for example, the appearance of holes on the rendered image, and the occurrence of depth discontinuity on the surface of the object at virtual image plane. Depth discontinuity causes artifacts on the rendered image. In this paper, these problems are solved by reconstructing disparity information at virtual camera position from neighboring two real cameras. In the experiments, high quality arbitrary viewpoint images were obtained.

  • PDF

Volume Rendering Using Multi-Textures (Multi-Textures를 이용한 Volume Rendering)

  • 박재영;이병일;최흥국
    • Proceedings of the Korea Institute of Convergence Signal Processing
    • /
    • 2000.12a
    • /
    • pp.169-172
    • /
    • 2000
  • Direct volume rendering has yet been restricted to high-end graphic workstations and special-purpose hardware, due to the large amount of trilinear interpolation, that are necessary to obtain high image quality. In this paper, we implemented the volume rendering techniques using the 2D-texture at the environment of standard PC hardware. In addition, we show how multi-texturing capabilities of modern PC graphics board are enable to volume rendering. Besides using extended OpenGL function, we improved pixel operations and rendering capacity.

  • PDF

Real-Time Haptic Rendering of Slowly Deformable Bodies Based on Two Dimensional Visual Information for Telemanipulation (원격조작을 위한 2차원 영상정보에 기반한 저속 변형체의 실시간 햅틱 렌더링)

  • Kim, Jung-Sik;Kim, Young-Jin;Kim, Jung
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.31 no.8
    • /
    • pp.855-861
    • /
    • 2007
  • Haptic rendering is a process providing force feedback during interactions between a user and a virtual object. This paper presents a real-time haptic rendering technique for deformable objects based on visual information of intervention between a tool and a real object in a remote place. A user can feel the artificial reaction force through a haptic device in real-time when a slave system exerts manipulation tasks on a deformable object. The models of the deformable object and the manipulator are created from the captured image obtained with a CCD camera and the recognition of objects is achieved using image processing techniques. The force at a rate of 1 kHz for stable haptic interaction is deduced using extrapolation of forces at a low update rate. The rendering algorithm developed was tested and validated on a test platform consisting of a one-dimensional indentation device and an off-the shelf force feedback device. This software system can be used in a cellular manipulation system providing artificial force feedback to enhance a success rate of operations.

Omnidirectional Camera-based Image Rendering Synchronization System Using Head Mounted Display (헤드마운티드 디스플레이를 활용한 전방위 카메라 기반 영상 렌더링 동기화 시스템)

  • Lee, Seungjoon;Kang, Suk-Ju
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.67 no.6
    • /
    • pp.782-788
    • /
    • 2018
  • This paper proposes a novel method for the omnidirectional camera-based image rendering synchronization system using head mounted display. There are two main processes in the proposed system. The first one is rendering 360-degree images which are remotely photographed to head mounted display. This method is based on transmission control protocol/internet protocol(TCP/IP), and the sequential images are rapidly captured and transmitted to the server using TCP/IP protocol with the byte array data format. Then, the server collects the byte array data, and make them into images. Finally, the observer can see them while wearing head mounted display. The second process is displaying the specific region by detecting the user's head rotation. After extracting the user's head Euler angles from head mounted display's inertial measurement units sensor, the proposed system display the region based on these angles. In the experimental results, rendering the original image at the same resolution in a given network environment causes loss of frame rate, and rendering at the same frame rate results in loss of resolution. Therefore, it is necessary to select optimal parameters considering environmental requirements.

Reconstruction of Color-Volume Data for Three-Dimensional Human Anatomic Atlas (3차원 인체 해부도 작성을 위한 칼라 볼륨 데이터의 입체 영상 재구성)

  • 김보형;이철희
    • Journal of Biomedical Engineering Research
    • /
    • v.19 no.2
    • /
    • pp.199-210
    • /
    • 1998
  • In this paper, we present a 3D reconstruction method of color volume data for a computerized human atlas. Binary volume rendering which takes the advantages of object-order ray traversal and run-length encoding visualizes 3D organs at an interactive speed in a general PC without the help of specific hardwares. This rendering method improves the rendering speed by simplifying the determination of the pixel value of an intermediate depth image and applying newly developed normal vector calculation method. Moreover, we describe the 3D boundary encoding that reduces the involved data considerably without the penalty of image quality. The interactive speed of the binary rendering and the storage efficiency of 3D boundary encoding will accelerate the development of the PC-based human atlas.

  • PDF

Occlusion-based Direct Volume Rendering for Computed Tomography Image

  • Jung, Younhyun
    • Journal of Multimedia Information System
    • /
    • v.5 no.1
    • /
    • pp.35-42
    • /
    • 2018
  • Direct volume rendering (DVR) is an important 3D visualization method for medical images as it depicts the full volumetric data. However, because DVR renders the whole volume, regions of interests (ROIs) such as a tumor that are embedded within the volume maybe occluded from view. Thus, conventional 2D cross-sectional views are still widely used, while the advantages of the DVR are often neglected. In this study, we propose a new visualization algorithm where we augment the 2D slice of interest (SOI) from an image volume with volumetric information derived from the DVR of the same volume. Our occlusion-based DVR augmentation for SOI (ODAS) uses the occlusion information derived from the voxels in front of the SOI to calculate a depth parameter that controls the amount of DVR visibility which is used to provide 3D spatial cues while not impairing the visibility of the SOI. We outline the capabilities of our ODAS and through a variety of computer tomography (CT) medical image examples, compare it to a conventional fusion of the SOI and the clipped DVR.

Rendering Trees Using Billboarding Method with View Dependent Texture Re-creation in Real-Time (시점 종속적 빌보드 텍스쳐 재생성을 이용한 나무의 실시간 렌더링)

  • Min, Sung-Hwan;Kim, Chang-Hun
    • Journal of the Korea Computer Graphics Society
    • /
    • v.8 no.4
    • /
    • pp.9-16
    • /
    • 2002
  • In this paper we propose an extended billboarding method for rendering trees. It's difficult circumstance that PC renders forest by general polygon rendering in real-time, because tree object consists of tremendous leaves and limbs. Our method re-creates an appropriate texture image by 3D image warping equation each frame, and we use it for a billboarding method. For speed up, we use warping method with a image pyramid and image caching.

  • PDF

Volume Rendering Architecture of Mobile Medical Image using Cloud Computing (클라우드 컴퓨팅을 활용한 모바일 의료영상 볼륨렌더링 아키텍처)

  • Lee, Woongkyu;Nam, Doohee
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.14 no.4
    • /
    • pp.101-106
    • /
    • 2014
  • The era came that by having fastest internet and smart phone makes cloud computing really a big merit. This paper proposes architecture for medical image volume rendering in mobile environment using cloud computing. This architecture to replace expensive workstation server and storage it use one of the service of cloud computing IaaS(Infrastructure as a Service). And this paper propose to use webGL to get rid of restriction of mobile hardware. By this research, it is expected that medical image volume rendering service in mobile environment is more effective and can be a foundation work.

Effects of Depth Map Quantization for Computer-Generated Multiview Images using Depth Image-Based Rendering

  • Kim, Min-Young;Cho, Yong-Joo;Choo, Hyon-Gon;Kim, Jin-Woong;Park, Kyoung-Shin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.5 no.11
    • /
    • pp.2175-2190
    • /
    • 2011
  • This paper presents the effects of depth map quantization for multiview intermediate image generation using depth image-based rendering (DIBR). DIBR synthesizes multiple virtual views of a 3D scene from a 2D image and its associated depth map. However, it needs precise depth information in order to generate reliable and accurate intermediate view images for use in multiview 3D display systems. Previous work has extensively studied the pre-processing of the depth map, but little is known about depth map quantization. In this paper, we conduct an experiment to estimate the depth map quantization that affords acceptable image quality to generate DIBR-based multiview intermediate images. The experiment uses computer-generated 3D scenes, in which the multiview images captured directly from the scene are compared to the multiview intermediate images constructed by DIBR with a number of quantized depth maps. The results showed that there was no significant effect on depth map quantization from 16-bit to 7-bit (and more specifically 96-scale) on DIBR. Hence, a depth map above 7-bit is needed to maintain sufficient image quality for a DIBR-based multiview 3D system.