• Title/Summary/Keyword: RGB-D images

Search Result 109, Processing Time 0.029 seconds

Convenient View Calibration of Multiple RGB-D Cameras Using a Spherical Object (구형 물체를 이용한 다중 RGB-D 카메라의 간편한 시점보정)

  • Park, Soon-Yong;Choi, Sung-In
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.8
    • /
    • pp.309-314
    • /
    • 2014
  • To generate a complete 3D model from depth images of multiple RGB-D cameras, it is necessary to find 3D transformations between RGB-D cameras. This paper proposes a convenient view calibration technique using a spherical object. Conventional view calibration methods use either planar checkerboards or 3D objects with coded-pattern. In these conventional methods, detection and matching of pattern features and codes takes a significant time. In this paper, we propose a convenient view calibration method using both 3D depth and 2D texture images of a spherical object simultaneously. First, while moving the spherical object freely in the modeling space, depth and texture images of the object are acquired from all RGB-D camera simultaneously. Then, the external parameters of each RGB-D camera is calibrated so that the coordinates of the sphere center coincide in the world coordinate system.

Synthesis method of elemental images from Kinect images for space 3D image (공간 3D 영상디스플레이를 위한 Kinect 영상의 요소 영상 변환방법)

  • Ryu, Tae-Kyung;Hong, Seok-Min;Kim, Kyoung-Won;Lee, Byung-Gook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2012.05a
    • /
    • pp.162-163
    • /
    • 2012
  • In this paper, we propose a synthesis method of elemental images from Kinect images for 3D integral imaging display. Since RGB images and depth image obtained from Kinect are not able to display 3D images in integral imaging system, we need transform the elemental images in integral imaging display. To do so, we synthesize the elemental images based on the geometric optics mapping from the depth plane images obtained from RGB image and depth image. To show the usefulness of the proposed system, we carry out the preliminary experiments using the two person object and present the experimental results.

  • PDF

Prediction of concrete slump by RGB-D image feature fusion

  • Huansen Chen;Jianhong Yang;Huaiying Fang;Shaojie Wu;Bohong Lin
    • Computers and Concrete
    • /
    • v.34 no.5
    • /
    • pp.535-546
    • /
    • 2024
  • Slump is an important index for concrete fluidity, which has a direct guiding effect on construction. In recent years, using RGB images for evaluating slump has been confirmed by scholars. Based on previous studies, this paper investigates the superiority of RGB-D image data over RGB image data in predicting slump of concrete and proposes three RGB-D fusion models: The early-stage-fusion model performs feature fusion in the data input stage, while the fully-connected-layer-fusion model performs feature fusion in the classification layer and the middle-stage-fusion model performs feature fusion after each residual block. In the classification of slump 120 mm, 150 mm and 200 mm, the Precision, Recall and F1-score are used to evaluate the model's ability to classify a single class, and the Accuracy, Macro-F1, Kappa and MCC are used to evaluate the model's performance. The experimental results showed that compared with the model using only RGB images, the fusion model achieve better performance, indicating that RGB-D image data can better evaluate concrete slump.

A method of improving the quality of 3D images acquired from RGB-depth camera (깊이 영상 카메라로부터 획득된 3D 영상의 품질 향상 방법)

  • Park, Byung-Seo;Kim, Dong-Wook;Seo, Young-Ho
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.5
    • /
    • pp.637-644
    • /
    • 2021
  • In general, in the fields of computer vision, robotics, and augmented reality, the importance of 3D space and 3D object detection and recognition technology has emerged. In particular, since it is possible to acquire RGB images and depth images in real time through an image sensor using Microsoft Kinect method, many changes have been made to object detection, tracking and recognition studies. In this paper, we propose a method to improve the quality of 3D reconstructed images by processing images acquired through a depth-based (RGB-Depth) camera on a multi-view camera system. In this paper, a method of removing noise outside an object by applying a mask acquired from a color image and a method of applying a combined filtering operation to obtain the difference in depth information between pixels inside the object is proposed. Through each experiment result, it was confirmed that the proposed method can effectively remove noise and improve the quality of 3D reconstructed image.

Stream Environment Monitoring using UAV Images (RGB, Thermal Infrared) (UAV 영상(RGB, 적외 열 영상)을 활용한 하천환경 모니터링)

  • Kang, Joon-Oh;Kim, Dal-Joo;Han, Woong-Ji;Lee, Yong-Chang
    • Journal of Urban Science
    • /
    • v.6 no.2
    • /
    • pp.17-27
    • /
    • 2017
  • Recently, civil complaints have increased due to water pollution and bad smell in rivers. Therefore, attention is focused on improving the river environment. The purpose of this study is to acquire RGB and thermal infrared images using UAV for sewage outlet and to monitor the status of stream pollution and the applicability UAV based images for river embankment maintenance plan was examined. The accuracy of the 3D model was examination by SfM(Structure from Motion) based images analysis on river embankment maintenance area. Especially, The wastewater discharged from the factory near the river was detected as an thermal infrared images and the flow of wastewater was monitored. As a result of the study, we could monitor the cause and flows of wastewater pollution by detecting temperature change caused by wastewater inflow using UAV images. In addition, UAV based a high precision 3D model (DTM, Digital Topographic Map, Orthophoto Mosaic) was produced to obtain precise DSM(Digital Surface Model) and vegetation cover information for river embankment maintenance.

  • PDF

Information Hiding Method based on Interpolation using Max Difference of RGB Pixel for Color Images (컬러 영상의 RGB 화소 최대차분 기반 보간법을 이용한 정보은닉 기법)

  • Lee, Joon-Ho;Kim, Pyung-Han;Jung, Ki-Hyun;Yoo, Kee-Young
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.4
    • /
    • pp.629-639
    • /
    • 2017
  • Interpolation based information hiding methods are widely used to get information security. Conventional interpolation methods use the neighboring pixel value and simple calculation like average to embed secret bit stream into the image. But these information hiding methods are not appropriate to color images like military images because the characteristics of military images are not considered and these methods are restricted in grayscale images. In this paper, the new information hiding method based on interpolation using RGB pixel values of color image is proposed and the effectiveness is analyzed through experiments.

Quality Enhancement of 3D Volumetric Contents Based on 6DoF for 5G Telepresence Service

  • Byung-Seo Park;Woosuk Kim;Jin-Kyum Kim;Dong-Wook Kim;Young-Ho Seo
    • Journal of Web Engineering
    • /
    • v.21 no.3
    • /
    • pp.729-750
    • /
    • 2022
  • In general, the importance of 6DoF (degree of freedom) 3D (dimension) volumetric contents technology is emerging in 5G (generation) telepresence service, Web-based (WebGL) graphics, computer vision, robotics, and next-generation augmented reality. Since it is possible to acquire RGB images and depth images in real-time through depth sensors that use various depth acquisition methods such as time of flight (ToF) and lidar, many changes have been made in object detection, tracking, and recognition research. In this paper, we propose a method to improve the quality of 3D models for 5G telepresence by processing images acquired through depth and RGB cameras on a multi-view camera system. In this paper, the quality is improved in two major ways. The first concerns the shape of the 3D model. A method of removing noise outside the object by applying a mask obtained from a color image and a combined filtering operation to obtain the difference in depth information between pixels inside the object were proposed. Second, we propose an illumination compensation method for images acquired through a multi-view camera system for photo-realistic 3D model generation. It is assumed that the three-dimensional volumetric shooting is done indoors, and the location and intensity of illumination according to time are constant. Since the multi-view camera uses a total of 8 pairs and converges toward the center of space, the intensity and angle of light incident on each camera are different even if the illumination is constant. Therefore, all cameras take a color correction chart and use a color optimization function to obtain a color conversion matrix that defines the relationship between the eight acquired images. Using this, the image input from all cameras is corrected based on the color correction chart. It was confirmed that the quality of the 3D model could be improved by effectively removing noise due to the proposed method when acquiring images of a 3D volumetric object using eight cameras. It has been experimentally proven that the color difference between images is reduced.

A Novel RGB Channel Assimilation for Hyperspectral Image Classification using 3D-Convolutional Neural Network with Bi-Long Short-Term Memory

  • M. Preethi;C. Velayutham;S. Arumugaperumal
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.3
    • /
    • pp.177-186
    • /
    • 2023
  • Hyperspectral imaging technology is one of the most efficient and fast-growing technologies in recent years. Hyperspectral image (HSI) comprises contiguous spectral bands for every pixel that is used to detect the object with significant accuracy and details. HSI contains high dimensionality of spectral information which is not easy to classify every pixel. To confront the problem, we propose a novel RGB channel Assimilation for classification methods. The color features are extracted by using chromaticity computation. Additionally, this work discusses the classification of hyperspectral image based on Domain Transform Interpolated Convolution Filter (DTICF) and 3D-CNN with Bi-directional-Long Short Term Memory (Bi-LSTM). There are three steps for the proposed techniques: First, HSI data is converted to RGB images with spatial features. Before using the DTICF, the RGB images of HSI and patch of the input image from raw HSI are integrated. Afterward, the pair features of spectral and spatial are excerpted using DTICF from integrated HSI. Those obtained spatial and spectral features are finally given into the designed 3D-CNN with Bi-LSTM framework. In the second step, the excerpted color features are classified by 2D-CNN. The probabilistic classification map of 3D-CNN-Bi-LSTM, and 2D-CNN are fused. In the last step, additionally, Markov Random Field (MRF) is utilized for improving the fused probabilistic classification map efficiently. Based on the experimental results, two different hyperspectral images prove that novel RGB channel assimilation of DTICF-3D-CNN-Bi-LSTM approach is more important and provides good classification results compared to other classification approaches.

Robust 2D human upper-body pose estimation with fully convolutional network

  • Lee, Seunghee;Koo, Jungmo;Kim, Jinki;Myung, Hyun
    • Advances in robotics research
    • /
    • v.2 no.2
    • /
    • pp.129-140
    • /
    • 2018
  • With the increasing demand for the development of human pose estimation, such as human-computer interaction and human activity recognition, there have been numerous approaches to detect the 2D poses of people in images more efficiently. Despite many years of human pose estimation research, the estimation of human poses with images remains difficult to produce satisfactory results. In this study, we propose a robust 2D human body pose estimation method using an RGB camera sensor. Our pose estimation method is efficient and cost-effective since the use of RGB camera sensor is economically beneficial compared to more commonly used high-priced sensors. For the estimation of upper-body joint positions, semantic segmentation with a fully convolutional network was exploited. From acquired RGB images, joint heatmaps accurately estimate the coordinates of the location of each joint. The network architecture was designed to learn and detect the locations of joints via the sequential prediction processing method. Our proposed method was tested and validated for efficient estimation of the human upper-body pose. The obtained results reveal the potential of a simple RGB camera sensor for human pose estimation applications.

Spatial-temporal texture features for 3D human activity recognition using laser-based RGB-D videos

  • Ming, Yue;Wang, Guangchao;Hong, Xiaopeng
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.3
    • /
    • pp.1595-1613
    • /
    • 2017
  • The IR camera and laser-based IR projector provide an effective solution for real-time collection of moving targets in RGB-D videos. Different from the traditional RGB videos, the captured depth videos are not affected by the illumination variation. In this paper, we propose a novel feature extraction framework to describe human activities based on the above optical video capturing method, namely spatial-temporal texture features for 3D human activity recognition. Spatial-temporal texture feature with depth information is insensitive to illumination and occlusions, and efficient for fine-motion description. The framework of our proposed algorithm begins with video acquisition based on laser projection, video preprocessing with visual background extraction and obtains spatial-temporal key images. Then, the texture features encoded from key images are used to generate discriminative features for human activity information. The experimental results based on the different databases and practical scenarios demonstrate the effectiveness of our proposed algorithm for the large-scale data sets.