• Title/Summary/Keyword: RGB-D images

Search Result 109, Processing Time 0.024 seconds

A New Method for Color Feature Representation of Color Image in Content-Based Image Retrieval - 2D Projection Maps

  • Ha, Seok-Wun
    • Journal of information and communication convergence engineering
    • /
    • v.2 no.2
    • /
    • pp.123-127
    • /
    • 2004
  • The most popular technique for image retrieval in a heterogeneous collection of color images is the comparison of images based on their color histogram. The color histogram describes the distribution of colors in the color space of a color image. In the most image retrieval systems, the color histogram is used to compute similarities between the query image and all the images in a database. But, small changes in the resolution, scaling, and illumination may cause important modifications of the color histogram, and so two color images may be considered to be very different from each other even though they have completely related semantics. A new method of color feature representation based on the 3-dimensional RGB color map is proposed to improve the defects of the color histogram. The proposed method is based on the three 2-dimensional projection map evaluated by projecting the RGB color space on the RG, GB, and BR surfaces. The experimental results reveal that the proposed is less sensitive to small changes in the scene and that achieve higher retrieval performances than the traditional color histogram.

A Study on Improving Brightness Values for Real Representations of Material Images in Lightscape (Lightscape에서의 사실적 재질 이미지 표현을 위한 Brightness 수치 개선에 관한 연구)

  • Chang, Jun-Ho;Park, Ji-Ae;Choi, An-Seop
    • Journal of the Korean Institute of Illuminating and Electrical Installation Engineers
    • /
    • v.21 no.2
    • /
    • pp.1-8
    • /
    • 2007
  • This study is to generate optimal values of brightness and achieve more realistic images by varying such values according to individual materials in order to create better quality simulation images using Lightscape. Because there are discolorations of materials by adding the radiation of luminaires to that of daylight during the daytime, when low brightness values are inputted in case of daytime, the results are founded that images of materials can be seen much similar with real images. And in most of the materials, when low values in the daytime and high values in the nighttime are inputted this study verifies that the realest simulation images can be obtained using Lightscape, because the RGB data of simulation images are almost similar with that of real materials.

Noise Reduction Method Using Randomized Unscented Kalman Filter for RGB+D Camera Sensors (랜덤 무향 칼만 필터를 이용한 RGB+D 카메라 센서의 잡음 보정 기법)

  • Kwon, Oh-Seol
    • Journal of Broadcast Engineering
    • /
    • v.25 no.5
    • /
    • pp.808-811
    • /
    • 2020
  • This paper proposes a method to minimize the error of the Kinect camera sensor by using a random undirected Kalman filter. Kinect cameras, which provide RGB values and depth information, cause nonlinear errors in the sensor, causing problems in various applications such as skeleton detection. Conventional methods have tried to remove errors by using various filtering techniques. However, there is a limit to removing nonlinear noise effectively. Therefore, in this paper, a randomized unscented Kalman filter was applied to predict and update the nonlinear noise characteristics, we next tried to enhance a performance of skeleton detection. The experimental results confirmed that the proposed method is superior to the conventional method in quantitative results and reconstructed images on 3D space.

Transformer-based dense 3D reconstruction from RGB images (RGB 이미지에서 트랜스포머 기반 고밀도 3D 재구성)

  • Xu, Jiajia;Gao, Rui;Wen, Mingyun;Cho, Kyungeun
    • Annual Conference of KIPS
    • /
    • 2022.11a
    • /
    • pp.646-647
    • /
    • 2022
  • Multiview stereo (MVS) 3D reconstruction of a scene from images is a fundamental computer vision problem that has been thoroughly researched in recent times. Traditionally, MVS approaches create dense correspondences by constructing regularizations and hand-crafted similarity metrics. Although these techniques have achieved excellent results in the best Lambertian conditions, traditional MVS algorithms still contain a lot of artifacts. Therefore, in this study, we suggest using a transformer network to accelerate the MVS reconstruction. The network is based on a transformer model and can extract dense features with 3D consistency and global context, which are necessary to provide accurate matching for MVS.

FBX Format Animation Generation System Combined with Joint Estimation Network using RGB Images (RGB 이미지를 이용한 관절 추정 네트워크와 결합된 FBX 형식 애니메이션 생성 시스템)

  • Lee, Yujin;Kim, Sangjoon;Park, Gooman
    • Journal of Broadcast Engineering
    • /
    • v.26 no.5
    • /
    • pp.519-532
    • /
    • 2021
  • Recently, in various fields such as games, movies, and animation, content that uses motion capture to build body models and create characters to express in 3D space is increasing. Studies are underway to generate animations using RGB-D cameras to compensate for problems such as the cost of cinematography in how to place joints by attaching markers, but the problem of pose estimation accuracy or equipment cost still exists. Therefore, in this paper, we propose a system that inputs RGB images into a joint estimation network and converts the results into 3D data to create FBX format animations in order to reduce the equipment cost required for animation creation and increase joint estimation accuracy. First, the two-dimensional joint is estimated for the RGB image, and the three-dimensional coordinates of the joint are estimated using this value. The result is converted to a quaternion, rotated, and an animation in FBX format is created. To measure the accuracy of the proposed method, the system operation was verified by comparing the error between the animation generated based on the 3D position of the marker by attaching a marker to the body and the animation generated by the proposed system.

Recognizing 3D Object's Attribute with Template Matching from RGB-D Images (RGB-D 영상으로부터 형판 정합을 이용한 3차원 물체의 속성 인식)

  • Kim, Dong-Ha;Kim, Joo-Hee;Im, Tae-Kwon;Kim, In-Cheol
    • Annual Conference of KIPS
    • /
    • 2015.04a
    • /
    • pp.766-769
    • /
    • 2015
  • 본 논문에서는 컬러 영상과 깊이 영상으로부터 영상 전체의 정보를 활용하는 형판 정합 방법으로 특징을 추출하여, 사물의 속성을 인식하는 시스템을 제안한다. 본 시스템은 입력 영상으로부터 더 많은 정보를 얻기 위해 컬러 영상과 깊이 영상을 함께 사용하였다. 그리고 영상의 부분적인 정보가 아닌 전체 정보를 활용하는 형판 정합 방법을 사용하여 속성 인식률을 향상 시켰다. 본 시스템의 성능을 확인하기 위해 워싱턴 대학에서 제공하는 RGB-D 데이터 집합을 이용하여 다른 특징들 및 분류기와 비교실험을 진행하였고, 본 논문에서 제안하는 시스템의 높은 성능을 확인할 수 있었다.

Linear Regression-based 1D Invariant Image for Shadow Detection and Removal in Single Natural Image (단일 자연 영상에서 그림자 검출 및 제거를 위한 선형 회귀 기반의 1D 불변 영상)

  • Park, Ki-Hong
    • Journal of Digital Contents Society
    • /
    • v.19 no.9
    • /
    • pp.1787-1793
    • /
    • 2018
  • Shadow is a common phenomenon observed in natural scenes, but it has a negative influence on image analysis such as object recognition, feature detection and scene analysis. Therefore, the process of detecting and removing shadows included in digital images must be considered as a pre-processing process of image analysis. In this paper, the existing methods for acquiring 1D invariant images, one of the feature elements for detecting and removing shadows contained in a single natural image, are described, and a method for obtaining 1D invariant images based on linear regression has been proposed. The proposed method calculates the log of the band-ratio between each channel of the RGB color image, and obtains the grayscale image line by linear regression. The final 1D invariant images were obtained by projecting the log image of the band-ratio onto the estimated grayscale image line. Experimental results show that the proposed method has lower computational complexity than the existing projection method using entropy minimization, and shadow detection and removal based on 1D invariant images are performed effectively.

Effective Multi-Modal Feature Fusion for 3D Semantic Segmentation with Multi-View Images (멀티-뷰 영상들을 활용하는 3차원 의미적 분할을 위한 효과적인 멀티-모달 특징 융합)

  • Hye-Lim Bae;Incheol Kim
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.12
    • /
    • pp.505-518
    • /
    • 2023
  • 3D point cloud semantic segmentation is a computer vision task that involves dividing the point cloud into different objects and regions by predicting the class label of each point. Existing 3D semantic segmentation models have some limitations in performing sufficient fusion of multi-modal features while ensuring both characteristics of 2D visual features extracted from RGB images and 3D geometric features extracted from point cloud. Therefore, in this paper, we propose MMCA-Net, a novel 3D semantic segmentation model using 2D-3D multi-modal features. The proposed model effectively fuses two heterogeneous 2D visual features and 3D geometric features by using an intermediate fusion strategy and a multi-modal cross attention-based fusion operation. Also, the proposed model extracts context-rich 3D geometric features from input point cloud consisting of irregularly distributed points by adopting PTv2 as 3D geometric encoder. In this paper, we conducted both quantitative and qualitative experiments with the benchmark dataset, ScanNetv2 in order to analyze the performance of the proposed model. In terms of the metric mIoU, the proposed model showed a 9.2% performance improvement over the PTv2 model using only 3D geometric features, and a 12.12% performance improvement over the MVPNet model using 2D-3D multi-modal features. As a result, we proved the effectiveness and usefulness of the proposed model.

A New Camera System Implementation for Realistic Media-based Contents (실감미디어 기반의 콘텐츠를 위한 카메라 시스템의 구현)

  • Seo, Young Ho;Lee, Yoon Hyuk;Koo, Ja Myung;Kim, Woo Youl;Kim, Bo Ra;Kim, Moon Seok;Kim, Dong Wook
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.9 no.2
    • /
    • pp.99-109
    • /
    • 2013
  • In this paper, we propose a new system which captures real depth and color information from natural scene and implemented it. Based on it, we produced stereo and multiview images for 3-dimensional stereoscopic contents and introduced the production of a digital hologram which is considered to the next-generation image. The system consists of both a camera system for capturing images which correspond to RGB and depth images and softwares (SWs) for various image processings which consist of pre-processing such as rectification and calibration, 3D warping, and computer generated hologram (CGH). The camera system use a vertical rig with two paris of depth and RGB camera and a specially manufactured cold mirror which has the different transmittance according to wavelength for obtaining images with the same view point. The wavelength of our mirror is about 850nm. Each algorithm was implemented using C and C++ and the implemented system can be operated in real-time.

3D Video Quality Improvement for 3D TV using Color Compensation (색상 보정을 통한 3차원 TV의 입체영상 화질 개선)

  • Jung, Kil-Soo;Kang, Min-Sung;Kim, Dong-Hyun;Sohn, Kwang-Hoon
    • Journal of Broadcast Engineering
    • /
    • v.15 no.6
    • /
    • pp.757-767
    • /
    • 2010
  • In this paper, we have studied the color compensation method for 3D that enables 3D color presentation similar to 2D. The color compensation method uses the difference of color presentation in 2D and 3D mode. First, the RGB I/O relationship curve was derived in 2D and 3D mode based on the input RGB color bar images. The relationship was modeled in modified power-law forms. Based on the modeling information, we generated color mapping tables, which can be used for compensating the difference of colors. The proposed color mapping block can be added at the output block of a 3DTV system, where the 2D content can be bypassed but the 3D content RGB data can be processed using the color mapping table. The experimental results show that the proposed method improves color presentation of a 3DTV system using a proper color compensation based on 2D presentation.