• Title/Summary/Keyword: Monocular Depth Map

Search Result 13, Processing Time 0.031 seconds

A Study of Depth Estimate using GPGPU in Monocular Image (GPGPU를 이용한 단일 영상에서의 깊이 추정에 관한 연구)

  • Yoo, Tae Hoon;Lee, Gang Seong;Park, Young Soo;Lee, Jong Yong;Lee, Sang Hun
    • Journal of Digital Convergence
    • /
    • v.11 no.12
    • /
    • pp.345-352
    • /
    • 2013
  • In this paper, a depth estimate method is proposed using GPU(Graphics Processing Unit) in monocular image. a monocular image is a 2D image with missing 3D depth information due to the camera projection and we used a monocular cue to recover the lost depth information by the projection present. The proposed algorithm uses an energy function which takes a variety of cues to create a more generalized and reliable depth map. But, a processing time is late because energy function is defined from the various monocular cues. Therefore, we propose a depth estimate method using GPGPU(General Purpose Graphics Processing Unit). The objective effectiveness of the algorithm is shown using PSNR(Peak Signal to Noise Ratio), a processing time is decrease by 61.22%.

Depth Map Extraction from the Single Image Using Pix2Pix Model (Pix2Pix 모델을 활용한 단일 영상의 깊이맵 추출)

  • Gang, Su Myung;Lee, Joon Jae
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.5
    • /
    • pp.547-557
    • /
    • 2019
  • To extract the depth map from a single image, a number of CNN-based deep learning methods have been performed in recent research. In this study, the GAN structure of Pix2Pix is maintained. this model allows to converge well, because it has the structure of the generator and the discriminator. But the convolution in this model takes a long time to compute. So we change the convolution form in the generator to a depthwise convolution to improve the speed while preserving the result. Thus, the seven down-sizing convolutional hidden layers in the generator U-Net are changed to depthwise convolution. This type of convolution decreases the number of parameters, and also speeds up computation time. The proposed model shows similar depth map prediction results as in the case of the existing structure, and the computation time in case of a inference is decreased by 64%.

Unsupervised Monocular Depth Estimation Using Self-Attention for Autonomous Driving (자율주행을 위한 Self-Attention 기반 비지도 단안 카메라 영상 깊이 추정)

  • Seung-Jun Hwang;Sung-Jun Park;Joong-Hwan Baek
    • Journal of Advanced Navigation Technology
    • /
    • v.27 no.2
    • /
    • pp.182-189
    • /
    • 2023
  • Depth estimation is a key technology in 3D map generation for autonomous driving of vehicles, robots, and drones. The existing sensor-based method has high accuracy but is expensive and has low resolution, while the camera-based method is more affordable with higher resolution. In this study, we propose self-attention-based unsupervised monocular depth estimation for UAV camera system. Self-Attention operation is applied to the network to improve the global feature extraction performance. In addition, we reduce the weight size of the self-attention operation for a low computational amount. The estimated depth and camera pose are transformed into point cloud. The point cloud is mapped into 3D map using the occupancy grid of Octree structure. The proposed network is evaluated using synthesized images and depth sequences from the Mid-Air dataset. Our network demonstrates a 7.69% reduction in error compared to prior studies.

Depth estimation and View Synthesis using Haze Information (실안개를 이용한 단일 영상으로부터의 깊이정보 획득 및 뷰 생성 알고리듬)

  • Soh, Yong-Seok;Hyun, Dae-Young;Lee, Sang-Uk
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2010.07a
    • /
    • pp.241-243
    • /
    • 2010
  • Previous approaches to the 2D to 3D conversion problem require heavy computation or considerable amount of user input. In this paper, we propose a rather simple method in estimating the depth map from a single image using a monocular depth cue: haze. Using the haze imaging model, we obtain the distance information and estimate a reliable depth map from a single scenery image. Using the depth map, we also suggest an algorithm that converts the single image to 3D stereoscopic images. We determine a disparity value for each pixel from the original 'left' image and generate a corresponding 'right' image. Results show that the algorithm gives well refined depth maps despite the simplicity of the approach.

  • PDF

High-Quality Depth Map Generation of Humans in Monocular Videos (단안 영상에서 인간 오브젝트의 고품질 깊이 정보 생성 방법)

  • Lee, Jungjin;Lee, Sangwoo;Park, Jongjin;Noh, Junyong
    • Journal of the Korea Computer Graphics Society
    • /
    • v.20 no.2
    • /
    • pp.1-11
    • /
    • 2014
  • The quality of 2D-to-3D conversion depends on the accuracy of the assigned depth to scene objects. Manual depth painting for given objects is labor intensive as each frame is painted. Specifically, a human is one of the most challenging objects for a high-quality conversion, as a human body is an articulated figure and has many degrees of freedom (DOF). In addition, various styles of clothes, accessories, and hair create a very complex silhouette around the 2D human object. We propose an efficient method to estimate visually pleasing depths of a human at every frame in a monocular video. First, a 3D template model is matched to a person in a monocular video with a small number of specified user correspondences. Our pose estimation with sequential joint angular constraints reproduces a various range of human motions (i.e., spine bending) by allowing the utilization of a fully skinned 3D model with a large number of joints and DOFs. The initial depth of the 2D object in the video is assigned from the matched results, and then propagated toward areas where the depth is missing to produce a complete depth map. For the effective handling of the complex silhouettes and appearances, we introduce a partial depth propagation method based on color segmentation to ensure the detail of the results. We compared the result and depth maps painted by experienced artists. The comparison shows that our method produces viable depth maps of humans in monocular videos efficiently.

Augmented Reality system Using Depth-map (Depth-Map을 이용한 객체 증강 시스템)

  • Ban, Kyeong-Jin;Kim, Jong-Chan;Kim, Kyoung-Ok;Kim, Eung-Kon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2010.10a
    • /
    • pp.343-344
    • /
    • 2010
  • markerless system to a two-dimensional imaging is used to estimate the depth map as a stereo vision system uses expensive equipment. We estimate the depth map from monocular image enhancement and object extracted relative to the vanishing point is estimated depth map. Augmented objects in order to get better virtual immersion depending on the distance of the objects should be drawn in different sizes. In this paper, creating images obtained from the vanishing point, and in-depth information on the augmented object, augmented with different sizes and improved engagement of inter-object interaction.

  • PDF

Relative Depth-Map Generation of Natural Scenes using Monocular Cues (단안단서를 이용한 자연영상의 상대적 깊이지도 생성)

  • Han Jong-Won;Jo Jin-Su;Lee Yill-Byung
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2006.06b
    • /
    • pp.367-369
    • /
    • 2006
  • 사람은 일반적으로 깊이를 지각하는데 두 눈으로 들어오는 영상의 시차(binocular disparity)를 이용하며 6-15m 정도의 범위 내에서는 매우 뛰어난 깊이 판별 능력을 보인다. 그러나 사람은 하나의 눈만으로도 깊이를 지각하는데 별 어려움을 느끼지 못한다. 이것은 공간의 깊이 지각 단서로 양안단서안이 아니라 다양한 단안단서(monocular Cue)들이 함께 사용되기 때문이다. 본 논문에서는 사람이 공간 깊이정보 파악에 사용하는 것으로 알려진 여러 단안 단서들 중 영상의 채도(saturation) 정보와 디포커스(defocus) 정보, 기하학적 깊이(geometric depth) 정보에 기반을 둔 단안 영상에서의 상대적 깊이지도의 생성방법을 제안한다.

  • PDF

Improved depth map generation method using Vanishing Point area (소실점 영역을 이용한 개선된 Depth-map 생성 기법)

  • Ban, Kyeong-Jin;Kim, Jong-Chan;Kim, Kyoung-Ok;Kim, Eung-Kon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2010.10a
    • /
    • pp.357-359
    • /
    • 2010
  • In monocular images that are used to determine the depth of the vanishing point, the buildings, roads and buildings, such as outdoor video or hallway with room inside for the interior structure, such as the vanishing point in the video is a very strong depth cue. Depth map using the vanishing point in the three-dimensional space, the two-dimensional imaging is used to restore the structure. But if there is a vanishing point vanishing point in the video also depends on the location of the relative depth of different ways to express that need. In this paper we present images of a vanishing point with respect to the improved depth-map was created. Proposed an area where the loss of seven points and areas defined as areas along the proposed direction of different depth.

  • PDF

Deep Learning-based Depth Map Estimation: A Review

  • Abdullah, Jan;Safran, Khan;Suyoung, Seo
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.1
    • /
    • pp.1-21
    • /
    • 2023
  • In this technically advanced era, we are surrounded by smartphones, computers, and cameras, which help us to store visual information in 2D image planes. However, such images lack 3D spatial information about the scene, which is very useful for scientists, surveyors, engineers, and even robots. To tackle such problems, depth maps are generated for respective image planes. Depth maps or depth images are single image metric which carries the information in three-dimensional axes, i.e., xyz coordinates, where z is the object's distance from camera axes. For many applications, including augmented reality, object tracking, segmentation, scene reconstruction, distance measurement, autonomous navigation, and autonomous driving, depth estimation is a fundamental task. Much of the work has been done to calculate depth maps. We reviewed the status of depth map estimation using different techniques from several papers, study areas, and models applied over the last 20 years. We surveyed different depth-mapping techniques based on traditional ways and newly developed deep-learning methods. The primary purpose of this study is to present a detailed review of the state-of-the-art traditional depth mapping techniques and recent deep learning methodologies. This study encompasses the critical points of each method from different perspectives, like datasets, procedures performed, types of algorithms, loss functions, and well-known evaluation metrics. Similarly, this paper also discusses the subdomains in each method, like supervised, unsupervised, and semi-supervised methods. We also elaborate on the challenges of different methods. At the conclusion of this study, we discussed new ideas for future research and studies in depth map research.

3D Environment Perception using Stereo Infrared Light Sources and a Camera (스테레오 적외선 조명 및 단일카메라를 이용한 3차원 환경인지)

  • Lee, Soo-Yong;Song, Jae-Bok
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.15 no.5
    • /
    • pp.519-524
    • /
    • 2009
  • This paper describes a new sensor system for 3D environment perception using stereo structured infrared light sources and a camera. Environment and obstacle sensing is the key issue for mobile robot localization and navigation. Laser scanners and infrared scanners cover $180^{\circ}$ and are accurate but too expensive. Those sensors use rotating light beams so that the range measurements are constrained on a plane. 3D measurements are much more useful in many ways for obstacle detection, map building and localization. Stereo vision is very common way of getting the depth information of 3D environment. However, it requires that the correspondence should be clearly identified and it also heavily depends on the light condition of the environment. Instead of using stereo camera, monocular camera and two projected infrared light sources are used in order to reduce the effects of the ambient light while getting 3D depth map. Modeling of the projected light pattern enabled precise estimation of the range. Two successive captures of the image with left and right infrared light projection provide several benefits, which include wider area of depth measurement, higher spatial resolution and the visibility perception.