• Title/Summary/Keyword: 3D방송

Search Result 1,332, Processing Time 0.025 seconds

Real-Time Augmented Reality on 3-D Mobile Display using Stereo Camera Tracking (스테레오 카메라 추적을 이용한 모바일 3차원 디스플레이 상의 실시간 증강현실)

  • Park, Jungsik;Seo, Byung-Kuk;Park, Jong-Il
    • Journal of Broadcast Engineering
    • /
    • v.18 no.3
    • /
    • pp.362-371
    • /
    • 2013
  • This paper presents a framework of real-time augmented reality on 3-D mobile display with stereo camera tracking. In the framework, camera poses are jointly estimated with the geometric relationship between stereoscopic images, which is based on model-based tracking. With the estimated camera poses, the virtual contents are correctly augmented on stereoscopic images through image rectification. For real-time performance, stereo camera tracking and image rectification are efficiently performed using multiple threads. Image rectification and color conversion are accelerated with a GPU processing. The proposed framework is tested and demonstrated on a commercial smartphone, which is equipped with a stereoscopic camera and a parallax barrier 3-D display.

A Web-based System for Embedding a Live Actor and Entity using X3DOM (X3DOM 을 이용한 라이브 행동자와 실체를 통합하기 위한 웹 기반 시스템)

  • Chheang, Vuthea;Ryu, Ga-Ae;Jeong, Sangkwon;Lee, Gookhwan;Yoo, Kwan-Hee
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2016.11a
    • /
    • pp.1-3
    • /
    • 2016
  • Mixed and augmented reality (MAR) refers to a spatially coordinated combination of media/information components that represent on the real world and its objects, and on the other those that are virtual, synthetic and computer generated including any combination of aural, visual and touch. The extensible 3D (X3D) is the ISO standard for defining 3D interactive web-based 3D content integrated with multimedia. In this paper, we propose a model to integrate live actor and entity that captured from Microsoft Kinect to be represented in Web-based mixed augmented reality world by using X3DOM by which X3D nodes can be integrated seamlessly into HTML5 DOM content.

  • PDF

3D Gaze Estimation and Interaction Technique (3차원 시선 추출 및 상호작용 기법)

  • Ki, Jeong-Seok;Jeon, Kyeong-Won;Kim, Sung-Kyu;Sohn, Kwang-Hoon;Kwon, Yong-Moo
    • Journal of Broadcast Engineering
    • /
    • v.11 no.4 s.33
    • /
    • pp.431-440
    • /
    • 2006
  • There are several researches on 2D gaze tracking techniques for the 2D screen for the Human-Computer Interaction. However, the researches for the gaze-based interaction to the stereo images or contents are not reported. The 3D display techniques are emerging now for the reality service. Moreover, the 3D interaction techniques are much more needed in the 3D contents service environments. This paper addresses gaze-based 3D interaction techniques on stereo display, such as parallax barrier or lenticular stereo display. This paper presents our researches on 3D gaze estimation and gaze-based interaction to stereo display.

3D Accuracy Analysis of Mobile Phone-based Stereo Images (모바일폰 기반 스테레오 영상에서 산출된 3차원 정보의 정확도 분석)

  • Ahn, Heeran;Kim, Jae-In;Kim, Taejung
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.677-686
    • /
    • 2014
  • This paper analyzes the 3D accuracy of stereo images captured from a mobile phone. For 3D accuracy evaluation, we have compared the accuracy result according to the amount of the convergence angle. In order to calculate the 3D model space coordinate of control points, we perform inner orientation, distortion correction and image geometry estimation. And the quantitative 3D accuracy was evaluated by transforming the 3D model space coordinate into the 3D object space coordinate. The result showed that relatively precise 3D information is generated in more than $17^{\circ}$ convergence angle. Consequently, it is necessary to set up stereo model structure consisting adequate convergence angle as an measurement distance and a baseline distance for accurate 3D information generation. It is expected that the result would be used to stereoscopic 3D contents and 3D reconstruction from images captured by a mobile phone camera.

A method of density scalability using SHVC codec in Video based Point Cloud Compression (SHVC 기반 V-PCC 3 차원 포인트 밀도 확장성 지원 방안)

  • Hwang, Yonghae;Kim, Junsik;Kim, Kyuheon
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.07a
    • /
    • pp.505-509
    • /
    • 2020
  • 포인트 클라우드 콘텐츠는 3 차원 공간에 수십만 개가 넘는 점들의 집합으로 이루어진 3D 데이터로 각 점들은 3 차원 공간의 좌표 데이터를 필요로 하고 추가적으로 색 (color), 반사율 (reflectance), 법선 벡터 (normal vector) 등과 같은 속성으로 구성되어 있다. 기존 2D 영상보다 한단계 높은 차원을 가진 3D 포인트 클라우드를 사용자에게 효율적으로 제공하기 위해서 고효율의 압축 기술 연구가 진행되고 있는데, 다양한 장치에서 발생하는 성능 차이에 구애 받지 않고 사용자에게 알맞은 서비스를 제공하기 위해서는 다양한 확장성에 대한 연구가 필요하다. 이에 본 논문에서는 포인트 클라우드 압축에 사용되는 Video-based Point Cloud Compression (V-PCC) 구조에 SHVC 코덱을 적용하여, 밀도 확장성을 갖는 포인트 클라우드 압축 비트스트림을 생성하는 방안을 제안하였다.

  • PDF

Three-dimensional beamforming techniques for LTE-A systems (LTE-A 시스템에서 3 차원 빔포밍 기법 연구)

  • Ji, Hyoungju;Shim, Byonghyo
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2015.11a
    • /
    • pp.43-44
    • /
    • 2015
  • LTE-Advanced system has been deployed with 2 and 4 transmission antennas (Tx) while the specification supports up to 8Tx. Due to deployment space, antenna dimension and complexity, the needs of deploying 8Tx system has not been motivated by operators. Recently, three dimensional (3D) beamforming with active antenna has attracted significant attention in the wireless industry. By incorporating 2D active array into LTE-A systems, the system offers freedom in controlling radiation on elevation and horizontal dimension. When the number of antennas increases in the form of 2D arrangement, spatial separation can be realized simultaneously in horizontal and elevation domain and vertical beam-steering can increase SINR of UEs in high floors. In this paper, we study the system operations and implementations for supporting 3D beamforming with 8Tx antennas. In our schemes, by reusing the conventional CSI feedback framework, the system can operate 2D active array without harming the backward compatibility. Evaluation results show that 3D beamforming provides capacity boosting over the conventional 2D beamforming systems while keeping same antenna structure.

  • PDF

UV Mapping Based Pose Estimation of Furniture Parts in Assembly Manuals (UV-map 기반의 신경망 학습을 이용한 조립 설명서에서의 부품의 자세 추정)

  • Kang, Isaac;Cho, Nam Ik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.07a
    • /
    • pp.667-670
    • /
    • 2020
  • 최근에는 증강현실, 로봇공학 등의 분야에서 객체의 위치 검출 이외에도, 객체의 자세에 대한 추정이 요구되고 있다. 객체의 자세 정보가 포함된 데이터셋은 위치 정보만 포함된 데이터셋에 비하여 상대적으로 매우 적기 때문에 인공 신경망 구조를 활용하기 어려운 측면이 있으나, 최근에 들어서는 기계학습 기반의 자세 추정 알고리즘들이 여럿 등장하고 있다. 본 논문에서는 이 가운데 Dense 6d Pose Object detector (DPOD) [11]의 구조를 기반으로 하여 가구의 조립 설명서에 그려진 가구 부품들의 자세를 추정하고자 한다. DPOD [11]는 입력으로 RGB 영상을 받으며, 해당 영상에서 자세를 추정하고자 하는 객체의 영역에 해당하는 픽셀들을 추정하고, 객체의 영역에 해당되는 각 픽셀에서 해당 객체의 3D 모델의 UV map 값을 추정한다. 이렇게 픽셀 개수만큼의 2D - 3D 대응이 생성된 이후에는, RANSAC과 PnP 알고리즘을 통해 RGB 영상에서의 객체와 객체의 3D 모델 간의 변환 관계 행렬이 구해지게 된다. 본 논문에서는 사전에 정해진 24개의 자세 후보들을 기반으로 가구 부품의 3D 모델을 2D에 투영한 RGB 영상들로 인공 신경망을 학습하였으며, 평가 시에는 실제 조립 설명서에서의 가구 부품의 자세를 추정하였다. 실험 결과 IKEA의 Stefan 의자 조립 설명서에 대하여 100%의 ADD score를 얻었으며, 추정 자세가 자세 후보군 중 정답 자세에 가장 근접한 경우를 정답으로 평가했을 때 100%의 정답률을 얻었다. 제안하는 신경망을 사용하였을 때, 가구 조립 설명서에서 가구 부품의 위치를 찾는 객체 검출기(object detection network)와, 각 개체의 종류를 구분하는 객체 리트리벌 네트워크(retrieval network)를 함께 사용하여 최종적으로 가구 부품의 자세를 추정할 수 있다.

  • PDF

3D Reenactment System of Soccer Game (3차원 축구 재연 시스템)

  • 이재호;김진우;김희정
    • Journal of Broadcast Engineering
    • /
    • v.8 no.1
    • /
    • pp.54-62
    • /
    • 2003
  • This paper presents a Soccer Game 3D Reencatment System which reenact the Important scene like getting a goal with image processing and computer graphics technologies. KBS Research Institute of Technology has developed the 3D Reenactment System of Soccer Game called ‘VPlay' to provide TV viewers with fresh images in soccer games. Vplay generates the reenactment of exciting and important soccer scenes by using computer graphics. Vplay extracts legion of players from video with color information, and then computes precise positions of players on the ground by using global motion estimation model and playground axis transformation model. The results are applied to locomotion generation module that generates the locomotion of virtual characters automatically. Using predefined motion and model library, Vplay reenacts the important scene in a quick and convenient manner Vplay was developed for live broadcasting of soccer games that demands rapid producing time and was used efficiently during past WorldCup and Asian Game.

Study on the technology of ATSC terrestrial DTV broadcasting service using single frequency networks (단일 주파수 망을 이용한 ATSC 지상파 디지털 TV 방송서비스 기술 연구)

  • 박성익;이용태;김승원;이수인
    • Journal of Broadcast Engineering
    • /
    • v.8 no.2
    • /
    • pp.109-115
    • /
    • 2003
  • In this paper we propose an efficient method to broadcast digital television signals using Single Frequency Networks (SFNs) in the Advanced Television Systems Committee (ATSC) transmission systems. Since the proposed schemes to synchronize multiple transmitters minimize the changes from the conventional ATSC system, the hardware complexity for the changes is very low. Our simulation results show that the proposed scheme makes less than 0.1 dB degradation at the threshold of visibility (TOV: BER= 3$\times$$10^{-6}$) in the additive white Gaussian noise (AWGN) channel. It is possible to reduce the performance degradation by increasing an initialization period of the proposed scheme.

Detecting Rectangular Image Regions in a Window Image for 3D Conversion (3D 변환을 위한 윈도우영상에서 사각 이미지 영역 검출)

  • Gil, Jong In;Lee, Jun Seok;Kim, Manbae
    • Journal of Broadcast Engineering
    • /
    • v.18 no.6
    • /
    • pp.795-807
    • /
    • 2013
  • In recent years, 2D-to-3D conversion techniques have gained much attraction. Most of conventional methods focused on natural images such as movie, animation and so forth. However, it is difficult to apply these techniques to window images mixed with text, image, logo, and icon. Also, different depth values of text pixels will cause distortion and a proper 3D image can not be delivered in some situations. To solve this problem, we propose a method to classify a given image into either a window or a natural image. For the window image, only rectangular image regions (RIR) are detected and converted in 3D. Other text and background are displayed in 2D. The proposed method was performed on more than 10,000 test images. In the experimental results, the detection ratio of window image reaches 97% and RIR detection ratio is 87%.