• Title/Summary/Keyword: 3D Cameras

Search Result 508, Processing Time 0.037 seconds

Design of Video Processor for Multi-View 3D Display (다시점 3차원 디스플레이용 비디오 프로세서의 설계)

  • 성준호;하태현;김성식;이성주;김재석
    • Journal of Broadcast Engineering
    • /
    • v.8 no.4
    • /
    • pp.452-464
    • /
    • 2003
  • In this paper, a multi-view 3D video processor was designed and implemented with several FPGAs for real-time applications. The 3D video processor receives 2D images from cameras (up to 16 cameras) and converts then to 3D video format for space-multiplexed 3D display. It can cope with various arrangements of 3D camera systems (or pixel arrays) and resolutions of 3D display. Tn order to verify the functions of 3D video Processor. some evaluation-board were made with five FPGAs.

Real-time Full-view 3D Human Reconstruction using Multiple RGB-D Cameras

  • Yoon, Bumsik;Choi, Kunwoo;Ra, Moonsu;Kim, Whoi-Yul
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.4 no.4
    • /
    • pp.224-230
    • /
    • 2015
  • This manuscript presents a real-time solution for 3D human body reconstruction with multiple RGB-D cameras. The proposed system uses four consumer RGB/Depth (RGB-D) cameras, each located at approximately $90^{\circ}$ from the next camera around a freely moving human body. A single mesh is constructed from the captured point clouds by iteratively removing the estimated overlapping regions from the boundary. A cell-based mesh construction algorithm is developed, recovering the 3D shape from various conditions, considering the direction of the camera and the mesh boundary. The proposed algorithm also allows problematic holes and/or occluded regions to be recovered from another view. Finally, calibrated RGB data is merged with the constructed mesh so it can be viewed from an arbitrary direction. The proposed algorithm is implemented with general-purpose computation on graphics processing unit (GPGPU) for real-time processing owing to its suitability for parallel processing.

Development of a Stereoscopic PTV (스테레오 PTV법의 개발)

  • Doh Deog Hee;Lee Won Je;Cho Yong Beom;Pyeon Yong Beom
    • Journal of the Korean Society of Visualization
    • /
    • v.1 no.1
    • /
    • pp.92-97
    • /
    • 2003
  • A new Stereoscopic PTV was developed using two CCD cameras, stereoscopic photogrammetry based on a 3D-PTV principle. Virtual images were produced for the benchmark test of the constructed Stereoscopic PTV technique. The arrangement of the two cameras was based on angular position. The calibration of cameras and the pair-matching of the three-dimensional velocity vectors were based on the Genetic Algorithm based 3D-PTV technique. The constructed Stereoscopic PTV technique was tested on the standard images of the impinged jet proposed by VSJ. The results obtained by the constructed system showed good agreements with the original data.

  • PDF

3D Positioning Accuracy Estimation of DMC in Compliance with Introducing High Resolution Digital Aerial Camera (고해상도 디지털항공사진 카메라 도입에 따른 DMC의 3차원 위치결정 정확도 평가)

  • Hahm, Chang-Hahk;Chang, Hwi-Jeong
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.27 no.1
    • /
    • pp.743-750
    • /
    • 2009
  • Since aerial photogrammetry by analog camera began in 1972, recently, high resolution digital camera is actively introduced to improve efficiency of aerial photogrammetry. This study investigated the 3D positioning accuracy of DMC(Digital Mapping Camera) among various high resolution aerial digital cameras to be developed for photogrammetry. For the research, we installed control points in test field around Incheon, and acquired analog and digital aerial photographs. By comparing 3D positioning accuracies of analog and digital photographs, there are few difference between two cameras, and the 3D positioning accuracies of two cameras was somewhat increased in case of aerotriangulation using additional control points based on GPS/IMU EO data.

An Adaptive Switching Mechanism for Three-Dimensional Hybrid Cameras (하이브리드 입체 카메라의 적응적인 스위칭 메커니즘)

  • Jang, Seok-Woo;Choi, Hyun-Jun;Lee, Suk-Yun;Huh, Moon-Haeng
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.14 no.3
    • /
    • pp.1459-1466
    • /
    • 2013
  • Recently, various types of three-dimensional cameras have been used to analyze surrounding environments. In this paper, we suggest a mechanism of adaptively switching active and passive cameras of hybrid cameras, which can extract 3D image information more accurately. The suggested method first obtains brightness and texture features representing the environment from input images. It then adaptively selects active and passive cameras by generating rules that reflect the extracted features. In experimental results, we show that a hybrid 3D camera consisting of passive and active cameras is set up and the proposed method can effectively choose appropriate cameras in the hybrid camera and make it possible to extract three dimensional information more accurately.

A study on comparison between 3D computer graphics cameras and actual cameras (3D컴퓨터그래픽스 가상현실 애니메이션 카메라와 실제카메라의 비교 연구 - Maya, Softimage 3D, XSI 소프트웨어와 실제 정사진과 동사진 카메라를 중심으로)

  • Kang, Chong-Jin
    • Cartoon and Animation Studies
    • /
    • s.6
    • /
    • pp.193-220
    • /
    • 2002
  • The world being made by computers showing great expanses and complex and various expression provides not simply communication places but also a new civilization and a new creative world. Among these, 3D computer graphics, 3D animation and virtual reality technology wore sublimated as a new culture and a new genre of art by joining graphic design and computer engineering. In this study, I tried to make a diagnosis of possibilities, limits and differences of expression in the area of virtual reality computer graphics animation as a comparison between camera action, angle of actual still camera and film camera and virtual software for 3D computer graphics software - Maya, XSI, Softimage3D.

  • PDF

Convenient View Calibration of Multiple RGB-D Cameras Using a Spherical Object (구형 물체를 이용한 다중 RGB-D 카메라의 간편한 시점보정)

  • Park, Soon-Yong;Choi, Sung-In
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.8
    • /
    • pp.309-314
    • /
    • 2014
  • To generate a complete 3D model from depth images of multiple RGB-D cameras, it is necessary to find 3D transformations between RGB-D cameras. This paper proposes a convenient view calibration technique using a spherical object. Conventional view calibration methods use either planar checkerboards or 3D objects with coded-pattern. In these conventional methods, detection and matching of pattern features and codes takes a significant time. In this paper, we propose a convenient view calibration method using both 3D depth and 2D texture images of a spherical object simultaneously. First, while moving the spherical object freely in the modeling space, depth and texture images of the object are acquired from all RGB-D camera simultaneously. Then, the external parameters of each RGB-D camera is calibrated so that the coordinates of the sphere center coincide in the world coordinate system.

CALOS : Camera And Laser for Odometry Sensing (CALOS : 주행계 추정을 위한 카메라와 레이저 융합)

  • Bok, Yun-Su;Hwang, Young-Bae;Kweon, In-So
    • The Journal of Korea Robotics Society
    • /
    • v.1 no.2
    • /
    • pp.180-187
    • /
    • 2006
  • This paper presents a new sensor system, CALOS, for motion estimation and 3D reconstruction. The 2D laser sensor provides accurate depth information of a plane, not the whole 3D structure. On the contrary, the CCD cameras provide the projected image of whole 3D scene, not the depth of the scene. To overcome the limitations, we combine these two types of sensors, the laser sensor and the CCD cameras. We develop a motion estimation scheme appropriate for this sensor system. In the proposed scheme, the motion between two frames is estimated by using three points among the scan data and their corresponding image points, and refined by non-linear optimization. We validate the accuracy of the proposed method by 3D reconstruction using real images. The results show that the proposed system can be a practical solution for motion estimation as well as for 3D reconstruction.

  • PDF

User classification and location tracking algorithm using deep learning (딥러닝을 이용한 사용자 구분 및 위치추적 알고리즘)

  • Park, Jung-tak;Lee, Sol;Park, Byung-Seo;Seo, Young-ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.78-79
    • /
    • 2022
  • In this paper, we propose a technique for tracking the classification and location of each user through body proportion analysis of the normalized skeletons of multiple users obtained using RGB-D cameras. To this end, each user's 3D skeleton is extracted from the 3D point cloud and body proportion information is stored. After that, the stored body proportion information is compared with the body proportion data output from the entire frame to propose a user classification and location tracking algorithm in the entire image.

  • PDF

Multiple Camera Calibration for Panoramic 3D Virtual Environment (파노라믹 3D가상 환경 생성을 위한 다수의 카메라 캘리브레이션)

  • 김세환;김기영;우운택
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.41 no.2
    • /
    • pp.137-148
    • /
    • 2004
  • In this paper, we propose a new camera calibration method for rotating multi-view cameras to generate image-based panoramic 3D Virtual Environment. Since calibration accuracy worsens with an increase in distance between camera and calibration pattern, conventional camera calibration algorithms are not proper for panoramic 3D VE generation. To remedy the problem, a geometric relationship among all lenses of a multi-view camera is used for intra-camera calibration. Another geometric relationship among multiple cameras is used for inter-camera calibration. First camera parameters for all lenses of each multi-view camera we obtained by applying Tsai's algorithm. In intra-camera calibration, the extrinsic parameters are compensated by iteratively reducing discrepancy between estimated and actual distances. Estimated distances are calculated using extrinsic parameters for every lens. Inter-camera calibration arranges multiple cameras in a geometric relationship. It exploits Iterative Closet Point (ICP) algorithm using back-projected 3D point clouds. Finally, by repeatedly applying intra/inter-camera calibration to all lenses of rotating multi-view cameras, we can obtain improved extrinsic parameters at every rotated position for a middle-range distance. Consequently, the proposed method can be applied to stitching of 3D point cloud for panoramic 3D VE generation. Moreover, it may be adopted in various 3D AR applications.