• Title/Summary/Keyword: 3-D video generation

Search Result 93, Processing Time 0.026 seconds

Stereoscopic Conversion based on Key Frames (키 프레임 기반 스테레오스코픽 변환 방법)

  • 김만배;박상훈
    • Journal of Broadcast Engineering
    • /
    • v.7 no.3
    • /
    • pp.219-228
    • /
    • 2002
  • In this paper, we propose a new method of converting 2D video into 3D stereoscopic video, called stereoscopic conversion. In general, stereoscopic images are produced using the motion informations. However unreliable motion informations obtained especially from block-based motion estimation cause the wrong generation of stereoscopic images. To solve for this problem, we propose a stereoscopic conversion method based upon the utilization of key frame that has the better accuracy of estimated motion informations. As well, as generation scheme of stereoscopic images associated with the motion type of each key frame is proposed. For the performance evaluation of our proposed method, we apply it to five test images and measure the accuracy of key frame-based stereoscopic conversion. Experimental results show that our proposed method has the accuracy more than about 90 percent in terms of the detection ratio of key frames.

2D Adjacency Matrix Generation using DCT for UWV Contents (DCT를 통한 UWV 콘텐츠의 2D 인접도 행렬 생성)

  • Xiaorui, Li;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.22 no.3
    • /
    • pp.366-374
    • /
    • 2017
  • Since a display device such as TV or digital signage is getting larger, the types of media is getting changed into wider view one such as UHD, panoramic and jigsaw-like media. Especially, panoramic and jigsaw-like media is realized by stitching video clips, which are captured by different camera or devices. However, a stitching process takes long time, and has difficulties in applying for a real-time process. Thus, this paper suggests to find out 2D Adjacency Matrix, which tells spatial relationships among those video clips in order to decrease a stitching processing time. Using the Discrete Cosine Transform (DCT), we convert the each frame of video source from the spatial domain (2D) into frequency domain. Based on the aforementioned features, 2D Adjacency Matrix of images could be found that we can efficiently make the spatial map of the images by using DCT. This paper proposes a new method of generating 2D adjacency matrix by using DCT for producing a panoramic and jigsaw-like media through various individual video clips.

Design of 3D Stereoscopic Electronic Book Authoring Tool Based on DirectX (DirectX기반 3차원 입체 eBook 영상 및 이미지 저작 도구 설계)

  • Park, Jinwoo;Lee, Keunhyoung;Kim, Jinmo;Hwang, Soyoung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2015.10a
    • /
    • pp.171-173
    • /
    • 2015
  • This paper proposes a design method of an authoring tool for making 3D e-book using DirectX development tools. There are several functions such as generation and modification of 3D objects, modification of textures, stereoscopic modes and pictures, video export and so on in the proposed authoring tool. To support these functions, we proposes design scheme such as data structures for generating 3D objects, anaglyph method using color differences and video export method using BandiCap library.

  • PDF

HEVC Encoder Optimization using Depth Information (깊이정보를 이용한 HEVC의 인코더 고속화 방법)

  • Lee, Yoon Jin;Bae, Dong In;Park, Gwang Hoon
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.640-655
    • /
    • 2014
  • Many of today's video systems have additional depth camera to provide extra features such as 3D support. Thanks to these changes made in multimedia system, it is now much easier to obtain depth information of the video. Depth information can be used in various areas such as object classification, background area recognition, and so on. With depth information, we can achieve even higher coding efficiency compared to only using conventional method. Thus, in this paper, we propose the 2D video coding algorithm which uses depth information on top of the next generation 2D video codec HEVC. Background area can be recognized with depth information and by performing HEVC with it, coding complexity can be reduced. If current CU is background area, we propose the following three methods, 1) Earlier stop split structure of CU with PU SKIP mode, 2) Limiting split structure of CU with CU information in temporal position, 3) Limiting the range of motion searching. We implement our proposal using HEVC HM 12.0 reference software. With these methods results shows that encoding complexity is reduced more than 40% with only 0.5% BD-Bitrate loss. Especially, in case of video acquired through the Kinect developed by Microsoft Corp., encoding complexity is reduced by max 53% without a loss of quality. So, it is expected that these techniques can apply real-time online communication, mobile or handheld video service and so on.

From Exoscope into the Next Generation

  • Nishiyama, Kenichi
    • Journal of Korean Neurosurgical Society
    • /
    • v.60 no.3
    • /
    • pp.289-293
    • /
    • 2017
  • An exoscope, high-definition video telescope operating monitor system to perform microsurgery has recently been proposed an alternative to the operating microscope. It enables surgeons to complete the operation assistance by visualizing magnified images on a display. The strong points of exoscope are the wide field of view and deep focus. It minimized the need for repositioning and refocusing during the procedure. On the other hand, limitation of magnifying object was an emphasizing weak point. The procedures are performed under 2D motion images with a visual perception through dynamic cue and stereoscopically viewing corresponding to the motion parallax. Nevertheless, stereopsis is required to improve hand and eye coordination for high precision works. Consequently novel 3D high-definition operating scopes with various mechanical designs have been developed according to recent high-tech innovations in a digital surgical technology. It will set the stage for the next generation in digital image based neurosurgery.

Fast Extraction of Objects of Interest from Images with Low Depth of Field

  • Kim, Chang-Ick;Park, Jung-Woo;Lee, Jae-Ho;Hwang, Jenq-Neng
    • ETRI Journal
    • /
    • v.29 no.3
    • /
    • pp.353-362
    • /
    • 2007
  • In this paper, we propose a novel unsupervised video object extraction algorithm for individual images or image sequences with low depth of field (DOF). Low DOF is a popular photographic technique which enables the representation of the photographer's intention by giving a clear focus only on an object of interest (OOI). We first describe a fast and efficient scheme for extracting OOIs from individual low-DOF images and then extend it to deal with image sequences with low DOF in the next part. The basic algorithm unfolds into three modules. In the first module, a higher-order statistics map, which represents the spatial distribution of the high-frequency components, is obtained from an input low-DOF image. The second module locates the block-based OOI for further processing. Using the block-based OOI, the final OOI is obtained with pixel-level accuracy. We also present an algorithm to extend the extraction scheme to image sequences with low DOF. The proposed system does not require any user assistance to determine the initial OOI. This is possible due to the use of low-DOF images. The experimental results indicate that the proposed algorithm can serve as an effective tool for applications, such as 2D to 3D and photo-realistic video scene generation.

  • PDF

An Objective No-Reference Perceptual Quality Assessment Metric based on Temporal Complexity and Disparity for Stereoscopic Video

  • Ha, Kwangsung;Bae, Sung-Ho;Kim, Munchurl
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.2 no.5
    • /
    • pp.255-265
    • /
    • 2013
  • 3DTV is expected to be a promising next-generation broadcasting service. On the other hand, the visual discomfort/fatigue problems caused by viewing 3D videos have become an important issue. This paper proposes a perceptual quality assessment metric for a stereoscopic video (SV-PQAM). To model the SV-PQAM, this paper presents the following features: temporal variance, disparity variation in intra-frames, disparity variation in inter-frames and disparity distribution of frame boundary areas, which affect the human perception of depth and visual discomfort for stereoscopic views. The four features were combined into the SV-PQAM, which then becomes a no-reference stereoscopic video quality perception model, as an objective quality assessment metric. The proposed SV-PQAM does not require a depth map but instead uses the disparity information by a simple estimation. The model parameters were estimated based on linear regression from the mean score opinion values obtained from the subjective perception quality assessments. The experimental results showed that the proposed SV-PQAM exhibits high consistency with subjective perception quality assessment results in terms of the Pearson correlation coefficient value of 0.808, and the prediction performance exhibited good consistency with a zero outlier ratio value.

  • PDF

Feature-Based Light and Shadow Estimation for Video Compositing and Editing (동영상 합성 및 편집을 위한 특징점 기반 조명 및 그림자 추정)

  • Hwang, Gyu-Hyun;Park, Sang-Hun
    • Journal of the Korea Computer Graphics Society
    • /
    • v.18 no.1
    • /
    • pp.1-9
    • /
    • 2012
  • Video-based modeling / rendering developed to produce photo-realistic video contents have been one of the important research topics in computer graphics and computer visions. To smoothly combine original input video clips and 3D graphic models, geometrical information of light sources and cameras used to capture a scene in the real world is essentially required. In this paper, we present a simple technique to estimate the position and orientation of an optimal light source from the topology of objects and the silhouettes of shadows appeared in the original video clips. The technique supports functions to generate well matched shadows as well as to render the inserted models by applying the estimated light sources. Shadows are known as an important visual cue that empirically indicates the relative location of objects in the 3D space. Thus our method can enhance realism in the final composed videos through the proposed shadow generation and rendering algorithms in real-time.

Generation of Stereoscopic Image from 2D Image based on Saliency and Edge Modeling (관심맵과 에지 모델링을 이용한 2D 영상의 3D 변환)

  • Kim, Manbae
    • Journal of Broadcast Engineering
    • /
    • v.20 no.3
    • /
    • pp.368-378
    • /
    • 2015
  • 3D conversion technology has been studied over past decades and integrated to commercial 3D displays and 3DTVs. The 3D conversion plays an important role in the augmented functionality of three-dimensional television (3DTV), because it can easily provide 3D contents. Generally, depth cues extracted from a static image is used for generating a depth map followed by DIBR (Depth Image Based Rendering) rendering for producing a stereoscopic image. However except some particular images, the existence of depth cues is rare so that the consistent quality of a depth map cannot be accordingly guaranteed. Therefore, it is imperative to make a 3D conversion method that produces satisfactory and consistent 3D for diverse video contents. From this viewpoint, this paper proposes a novel method with applicability to general types of image. For this, saliency as well as edge is utilized. To generate a depth map, geometric perspective, affinity model and binomic filter are used. In the experiments, the proposed method was performed on 24 video clips with a variety of contents. From a subjective test for 3D perception and visual fatigue, satisfactory and comfortable viewing of 3D contents was validated.

Realtime 3D Terrain Generation with Sound Data (사운드 데이터를 이용한 실시간 3차원 지형 생성)

  • Kim, Won-Seop;Chang, Kyu-Sik;Kim, Tae-Yong
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.184-189
    • /
    • 2008
  • In this paper, the sound data either from the sampled or streamed source are utilized for generating a map in the video game play for the dynamiccal use of sound data and synesthesia to users. When users can generate sound in real-time or put the sampled source, it is analyzed and re-processed through Fourier transformation to show the 3D map in dynamic shape over time. We interpolate the generated data to enable the game agents and objects to move.

  • PDF