• Title/Summary/Keyword: Depth Feature

Search Result 428, Processing Time 0.033 seconds

Fractal Depth Map Sequence Coding Algorithm with Motion-vector-field-based Motion Estimation

  • Zhu, Shiping;Zhao, Dongyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.1
    • /
    • pp.242-259
    • /
    • 2015
  • Three-dimensional video coding is one of the main challenges restricting the widespread applications of 3D video and free viewpoint video. In this paper, a novel fractal coding algorithm with motion-vector-field-based motion estimation for depth map sequence is proposed. We firstly add pre-search restriction to rule the improper domain blocks out of the matching search process so that the number of blocks involved in the search process can be restricted to a smaller size. Some improvements for motion estimation including initial search point prediction, threshold transition condition and early termination condition are made based on the feature of fractal coding. The motion-vector-field-based adaptive hexagon search algorithm on the basis of center-biased distribution characteristics of depth motion vector is proposed to accelerate the search. Experimental results show that the proposed algorithm can reach optimum levels of quality and save the coding time. The PSNR of synthesized view is increased by 0.56 dB with 36.97% bit rate decrease on average compared with H.264 Full Search. And the depth encoding time is saved by up to 66.47%. Moreover, the proposed fractal depth map sequence codec outperforms the recent alternative codecs by improving the H.264/AVC, especially in much bitrate saving and encoding time reduction.

3-D Recognition of Position using Epipolar Line and Matching from Stereo Image (두개의 영상으로부터 Epipolar Line과 Matching을 이용한 3차원 물체의 위치 인식)

  • Cho, Seok-Je;Park, Kil-Houm;Lee, Kwang-Ho;Kim, Young-Mo;Ha, Yeong-Ho
    • Proceedings of the KIEE Conference
    • /
    • 1987.07b
    • /
    • pp.1441-1444
    • /
    • 1987
  • Extraction of dept.h information from stereo image uses the matching process between them and this requires a lot of computational time. In this paper, a matching using the feature points on the epipolar line is presented to save the computations. Feature points are obtained in both image and correlated each other. With the coordinates of the matched feature points and camera geometry, the position and depth informations are identified.

  • PDF

SIFT-based Stereo Matching to Compensate Occluded Regions and Remove False Matching for 3D Reconstruction

  • Shin, Do-Kyung;Lee, Jeong-Ho;Moon, Young-Shik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.418-422
    • /
    • 2009
  • Generally, algorithms for generating disparity maps can be clssified into two categories: region-based method and feature-based method. The main focus of this research is to generate a disparity map with an accuracy depth information for 3-dimensional reconstructing. Basically, the region-based method and the feature-based method are simultaneously included in the proposed algorithm, so that the existing problems including false matching and occlusion can be effectively solved. As a region-based method, regions of false matching are extracted by the proposed MMAD(Modified Mean of Absolute Differences) algorithm which is a modification of the existing MAD(Mean of Absolute Differences) algorithm. As a feature-based method, the proposed method eliminates false matching errors by calculating the vector with SIFT and compensates the occluded regions by using a pair of adjacent SIFT matching points, so that the errors are reduced and the disparity map becomes more accurate.

  • PDF

Analogical Face Generation based on Feature Points

  • Yoon, Andy Kyung-yong;Park, Ki-cheul;Oh, Duck-kyo;Cho, Hye-young;Jang, Jung-hyuk
    • Journal of Multimedia Information System
    • /
    • v.6 no.1
    • /
    • pp.15-22
    • /
    • 2019
  • There are many ways to perform face recognition. The first step of face recognition is the face detection step. If the face is not found in the first step, the face recognition fails. Face detection research has many difficulties because it can be varied according to face size change, left and right rotation and up and down rotation, side face and front face, facial expression, and light condition. In this study, facial features are extracted and the extracted features are geometrically reconstructed in order to improve face recognition rate in extracted face region. Also, it is aimed to adjust face angle using reconstructed facial feature vector, and to improve recognition rate for each face angle. In the recognition attempt using the result after the geometric reconstruction, both the up and down and the left and right facial angles have improved recognition performance.

Bayesian-theory-based Fast CU Size and Mode Decision Algorithm for 3D-HEVC Depth Video Inter-coding

  • Chen, Fen;Liu, Sheng;Peng, Zongju;Hu, Qingqing;Jiang, Gangyi;Yu, Mei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.4
    • /
    • pp.1730-1747
    • /
    • 2018
  • Multi-view video plus depth (MVD) is a mainstream format of 3D scene representation in free viewpoint video systems. The advanced 3D extension of the high efficiency video coding (3D-HEVC) standard introduces new prediction tools to improve the coding performance of depth video. However, the depth video in 3D-HEVC is time consuming. To reduce the complexity of the depth video inter coding, we propose a fast coding unit (CU) size and mode decision algorithm. First, an off-line trained Bayesian model is built which the feature vector contains the depth levels of the corresponding spatial, temporal, and inter-component (texture-depth) neighboring largest CUs (LCUs). Then, the model is used to predict the depth level of the current LCU, and terminate the CU recursive splitting process. Finally, the CU mode search process is early terminated by making use of the mode correlation of spatial, inter-component (texture-depth), and inter-view neighboring CUs. Compared to the 3D-HEVC reference software HTM-10.0, the proposed algorithm reduces the encoding time of depth video and the total encoding time by 65.03% and 41.04% on average, respectively, with negligible quality degradation of the synthesized virtual view.

SuperDepthTransfer: Depth Extraction from Image Using Instance-Based Learning with Superpixels

  • Zhu, Yuesheng;Jiang, Yifeng;Huang, Zhuandi;Luo, Guibo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.10
    • /
    • pp.4968-4986
    • /
    • 2017
  • In this paper, we primarily address the difficulty of automatic generation of a plausible depth map from a single image in an unstructured environment. The aim is to extrapolate a depth map with a more correct, rich, and distinct depth order, which is both quantitatively accurate as well as visually pleasing. Our technique, which is fundamentally based on a preexisting DepthTransfer algorithm, transfers depth information at the level of superpixels. This occurs within a framework that replaces a pixel basis with one of instance-based learning. A vital superpixels feature enhancing matching precision is posterior incorporation of predictive semantic labels into the depth extraction procedure. Finally, a modified Cross Bilateral Filter is leveraged to augment the final depth field. For training and evaluation, experiments were conducted using the Make3D Range Image Dataset and vividly demonstrate that this depth estimation method outperforms state-of-the-art methods for the correlation coefficient metric, mean log10 error and root mean squared error, and achieves comparable performance for the average relative error metric in both efficacy and computational efficiency. This approach can be utilized to automatically convert 2D images into stereo for 3D visualization, producing anaglyph images that are visually superior in realism and simultaneously more immersive.

LiDAR Data Interpolation Algorithm for 3D-2D Motion Estimation (3D-2D 모션 추정을 위한 LiDAR 정보 보간 알고리즘)

  • Jeon, Hyun Ho;Ko, Yun Ho
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.12
    • /
    • pp.1865-1873
    • /
    • 2017
  • The feature-based visual SLAM requires 3D positions for the extracted feature points to perform 3D-2D motion estimation. LiDAR can provide reliable and accurate 3D position information with low computational burden, while stereo camera has the problem of the impossibility of stereo matching in simple texture image region, the inaccuracy in depth value due to error contained in intrinsic and extrinsic camera parameter, and the limited number of depth value restricted by permissible stereo disparity. However, the sparsity of LiDAR data may increase the inaccuracy of motion estimation and can even lead to the result of motion estimation failure. Therefore, in this paper, we propose three interpolation methods which can be applied to interpolate sparse LiDAR data. Simulation results obtained by applying these three methods to a visual odometry algorithm demonstrates that the selective bilinear interpolation shows better performance in the view point of computation speed and accuracy.

3D Model Retrieval based on Spherical Coordinate System (구면좌표계 기반에서 3차원 모델 검색)

  • Song, Ju-Whan;Choi, Seong-Hee
    • 전자공학회논문지 IE
    • /
    • v.46 no.1
    • /
    • pp.37-43
    • /
    • 2009
  • In this paper, we propose a new algorithm for 3D model retrieval based on spherical coordinate system. We obtains sample points in a polygons on 3D model. We convert a point in cartesian coordinates(x, y, z) to it in spherical coordinate. 3D shape features are achieved by adopting distribution of zenith of sample point in spherical coordinate. We used Osada's method for obtaining sample points on 3D model and the PCA method for the pose standardization 3D model. Princeton university's benchmark data was used for this research. Experimental results numerically show the precision improvement of proposed algorithm 12.6% in comparison with Vranic's depth buffer-based feature vector algorithm.

Dense RGB-D Map-Based Human Tracking and Activity Recognition using Skin Joints Features and Self-Organizing Map

  • Farooq, Adnan;Jalal, Ahmad;Kamal, Shaharyar
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.5
    • /
    • pp.1856-1869
    • /
    • 2015
  • This paper addresses the issues of 3D human activity detection, tracking and recognition from RGB-D video sequences using a feature structured framework. During human tracking and activity recognition, initially, dense depth images are captured using depth camera. In order to track human silhouettes, we considered spatial/temporal continuity, constraints of human motion information and compute centroids of each activity based on chain coding mechanism and centroids point extraction. In body skin joints features, we estimate human body skin color to identify human body parts (i.e., head, hands, and feet) likely to extract joint points information. These joints points are further processed as feature extraction process including distance position features and centroid distance features. Lastly, self-organized maps are used to recognize different activities. Experimental results demonstrate that the proposed method is reliable and efficient in recognizing human poses at different realistic scenes. The proposed system should be applicable to different consumer application systems such as healthcare system, video surveillance system and indoor monitoring systems which track and recognize different activities of multiple users.

An Object-based Stereo Matching Method Using Block-based Segmentation (블록 기반 영역 분할을 이용한 객체 기반 스테레오 정합 기법)

  • Kwak No-Yoon
    • Journal of Digital Contents Society
    • /
    • v.5 no.4
    • /
    • pp.257-263
    • /
    • 2004
  • This paper is related to the object-based stereo matching algorithm which makes it possible to estimate inner-region disparities for each segmented region. First, several sample points are selected for effectively representing the segmented region, Next, stereo matching is applied to the small area within segmented region which existed in the neighborhood or each sample point. Finally, inner-region disparities are interpolated using a plane equation with disparity of each selected sample. According to the proposed method, the problem of feature-based method that the depth estimation is possible only in the feature points can be solved through the propagation of the disparity in the sample point into the inside of the region. Also, as selecting sample points in contour of segmented region we can effectively suppress obscurity which is occurred in the depth estimation of the monotone region in area-based methods.

  • PDF