• Title/Summary/Keyword: Captured Image

Search Result 978, Processing Time 0.026 seconds

Distance Measurement Using the Kinect Sensor with Neuro-image Processing

  • Sharma, Kajal
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.4 no.6
    • /
    • pp.379-383
    • /
    • 2015
  • This paper presents an approach to detect object distance with the use of the recently developed low-cost Kinect sensor. The technique is based on Kinect color depth-image processing and can be used to design various computer-vision applications, such as object recognition, video surveillance, and autonomous path finding. The proposed technique uses keypoint feature detection in the Kinect depth image and advantages of depth pixels to directly obtain the feature distance in the depth images. This highly reduces the computational overhead and obtains the pixel distance in the Kinect captured images.

Investigation on the Applicability of Defocus Blur Variations to Depth Calculation Using Target Sheet Images Captured by a DSLR Camera

  • Seo, Suyoung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.38 no.2
    • /
    • pp.109-121
    • /
    • 2020
  • Depth calculation of objects in a scene from images is one of the most studied processes in the fields of image processing, computer vision, and photogrammetry. Conventionally, depth is calculated using a pair of overlapped images captured at different view points. However, there have been studies to calculate depths from a single image. Theoretically, it is known to be possible to calculate depth using the diameter of CoC (Circle of Confusion) caused by defocus under the assumption of a thin lens model. Thus, this study aims to verify the validity of the thin lens model to calculate depth from edge blur amount which corresponds to the radius of CoC. For this study, a commercially available DSLR (Digital Single Lens Reflex) camera was used to capture a set of target sheets which had different edge contrasts. In order to find out the pattern of the variations of edge blur against varying combination of FD (Focusing Distance) and OD (Object Distance), the camera was set to varying FD and target sheet images were captured at varying OD under each FD. Then, the edge blur and edge displacement were estimated from edge slope profiles using a brute-force method. The experimental results show that the pattern of the variations of edge blur observed in the target images was apart from their corresponding theoretical amounts derived under the thin lens assumption but can still be utilized to calculate depth from a single image for the cases similar to the limited conditions experimented under which the tendency between FD and OD is manifest.

Realtime 3D Human Full-Body Convergence Motion Capture using a Kinect Sensor (Kinect Sensor를 이용한 실시간 3D 인체 전신 융합 모션 캡처)

  • Kim, Sung-Ho
    • Journal of Digital Convergence
    • /
    • v.14 no.1
    • /
    • pp.189-194
    • /
    • 2016
  • Recently, there is increasing demand for image processing technology while activated the use of equipments such as camera, camcorder and CCTV. In particular, research and development related to 3D image technology using the depth camera such as Kinect sensor has been more activated. Kinect sensor is a high-performance camera that can acquire a 3D human skeleton structure via a RGB, skeleton and depth image in real-time frame-by-frame. In this paper, we develop a system. This system captures the motion of a 3D human skeleton structure using the Kinect sensor. And this system can be stored by selecting the motion file format as trc and bvh that is used for general purposes. The system also has a function that converts TRC motion captured format file into BVH format. Finally, this paper confirms visually through the motion capture data viewer that motion data captured using the Kinect sensor is captured correctly.

Crack Inspection and Mapping of Concrete Bridges using Integrated Image Processing Techniques (통합 이미지 처리 기술을 이용한 콘크리트 교량 균열 탐지 및 매핑)

  • Kim, Byunghyun;Cho, Soojin
    • Journal of the Korean Society of Safety
    • /
    • v.36 no.1
    • /
    • pp.18-25
    • /
    • 2021
  • In many developed countries, such as South Korea, efficiently maintaining the aging infrastructures is an important issue. Currently, inspectors visually inspect the infrastructure for maintenance needs, but this method is inefficient due to its high costs, long logistic times, and hazards to the inspectors. Thus, in this paper, a novel crack inspection approach for concrete bridges is proposed using integrated image processing techniques. The proposed approach consists of four steps: (1) training a deep learning model to automatically detect cracks on concrete bridges, (2) acquiring in-situ images using a drone, (3) generating orthomosaic images based on 3D modeling, and (4) detecting cracks on the orthmosaic image using the trained deep learning model. Cascade Mask R-CNN, a state-of-the-art instance segmentation deep learning model, was trained with 3235 crack images that included 2415 hard negative images. We selected the Tancheon overpass, located in Seoul, South Korea, as a testbed for the proposed approach, and we captured images of pier 34-37 and slab 34-36 using a commercial drone. Agisoft Metashape was utilized as a 3D model generation program to generate an orthomosaic of the captured images. We applied the proposed approach to four orthomosaic images that displayed the front, back, left, and right sides of pier 37. Using pixel-level precision referencing visual inspection of the captured images, we evaluated the trained Cascade Mask R-CNN's crack detection performance. At the coping of the front side of pier 37, the model obtained its best precision: 94.34%. It achieved an average precision of 72.93% for the orthomosaics of the four sides of the pier. The test results show that this proposed approach for crack detection can be a suitable alternative to the conventional visual inspection method.

The Compensation of Image Distortion on the X - Ray Image Intensifier (X - Ray 검사 시스템에서 Image Intensifier의 왜곡 보정)

  • Shin, Dong-Chul;Oh, Choon-Suk
    • The Transactions of the Korea Information Processing Society
    • /
    • v.5 no.4
    • /
    • pp.1044-1047
    • /
    • 1998
  • Non-destructive x-ray inspection system needs image intensifier to obtain the x-ray image. Captured image from image intensifier is distorted because the input plate of image intensifier has a spherical surface. In the research, in order to compensate this image distortion, we created the model of image intensifier and show mathematically that the image distortion was compensated. To show the performance of the proposed method, experiment was performed in real x-ray inspection system.

  • PDF

User Positioning Method Based on Image Similarity Comparison Using Single Camera (단일 카메라를 이용한 이미지 유사도 비교 기반의 사용자 위치추정)

  • Song, Jinseon;Hur, SooJung;Park, Yongwan;Choi, Jeonghee
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.8
    • /
    • pp.1655-1666
    • /
    • 2015
  • In this paper, user-position estimation method is proposed by using a single camera for both indoor and outdoor environments. Conventionally, the GPS of RF-based estimation methods have been widely studied in the literature for outdoor and indoor environments, respectively. Each method is useful only for indoor or outdoor environment. In this context, this study adopts a vision-based approach which can be commonly applicable to both environments. Since the distance or position cannot be extracted from a single still image, the reference images pro-stored in image database are used to identify the current position from the single still image captured by a single camera. The reference image is tagged with its captured position. To find the reference image which is the most similar to the current image, the SURF algorithm is used for feature extraction. The outliers in extracted features are discarded by using RANSAC algorithm. The performance of the proposed method is evaluated for two buildings and their outsides for both indoor and outdoor environments, respectively.

Generating a Stereoscopic Image from a Monoscopic Camera (단안 카메라를 이용한 입체영상 생성)

  • Lee, Dong-Woo;Lee, Kwan-Wook;Kim, Man-Bae
    • Journal of Broadcast Engineering
    • /
    • v.17 no.1
    • /
    • pp.17-25
    • /
    • 2012
  • In this paper, we propose a method of producing a stereoscopic image from multiple images captured from a monoscopic camera. By translating a camera in the horizontal direction, left and right images are chosen among N captured images. For this, image edges are extracted and a rotational angle is estimated from edge orientation. Also, a translational vector is also estimated from the correlation of projected image data. Then, two optimal images are chosen and subsequently compensated using the rotational angle as well as the translational vector in order to make a satisfactory stereoscopic image. The proposed method was performed on thirty-two test image set. The subjective visual fatigue test was carried out to validate the 3D quality of stereoscopic images. In terms of visual fatigue, the 3D satisfaction ratio reached approximately 84%.

Image Path Searching using Auto and Cross Correlations

  • Kim, Young-Bin;Ryu, Kwang-Ryol
    • Journal of information and communication convergence engineering
    • /
    • v.9 no.6
    • /
    • pp.747-752
    • /
    • 2011
  • The position detection of overlapping area in the interframe for image stitching using auto and cross correlation function (ACCF) and compounding one image with the stitching algorithm is presented in this paper. ACCF is used by autocorrelation to the featured area to extract the filter mask in the reference (previous) image and the comparing (current) image is used by crosscorrelation. The stitching is detected by the position of high correlation, and aligns and stitches the image in shifting the current image based on the moving vector. The ACCF technique results in a few computations and simplicity because the filter mask is given by the featuring block, and the position is enabled to detect a bit movement. Input image captured from CMOS is used to be compared with the performance between the ACCF and the window correlation. The results of ACCF show that there is no seam and distortion at the joint parts in the stitched image, and the detection performance of the moving vector is improved to 12% in comparison with the window correlation method.

Metadata Processing Technique for Similar Image Search of Mobile Platform

  • Seo, Jung-Hee
    • Journal of information and communication convergence engineering
    • /
    • v.19 no.1
    • /
    • pp.36-41
    • /
    • 2021
  • Text-based image retrieval is not only cumbersome as it requires the manual input of keywords by the user, but is also limited in the semantic approach of keywords. However, content-based image retrieval enables visual processing by a computer to solve the problems of text retrieval more fundamentally. Vision applications such as extraction and mapping of image characteristics, require the processing of a large amount of data in a mobile environment, rendering efficient power consumption difficult. Hence, an effective image retrieval method on mobile platforms is proposed herein. To provide the visual meaning of keywords to be inserted into images, the efficiency of image retrieval is improved by extracting keywords of exchangeable image file format metadata from images retrieved through a content-based similar image retrieval method and then adding automatic keywords to images captured on mobile devices. Additionally, users can manually add or modify keywords to the image metadata.

3D Image Capturing and 3D Content Generation for Realistic Broadcasting (실감방송을 위한 3차원 영상 촬영 및 3차원 콘텐츠 제작 기술)

  • Kang, Y.S.;Ho, Y.S.
    • Smart Media Journal
    • /
    • v.1 no.1
    • /
    • pp.10-16
    • /
    • 2012
  • Stereo and multi-view cameras have been used to capture the three-dimensional (3D) scene for 3D contents generation. Besides, depth sensors are frequently used to obtain 3D information of the captured scene in real time. In order to generate 3D contents from captured images, we need several preprocessing operations to reduce noises and distortions in the images. 3D contents are considered as the basic media for realistic broadcasting that provides photo-realistic and immersive feeling to users. In this paper, we show technical trends of 3D image capturing and contents generation, and explain some core techniques for 3D image processing for realistic 3DTV broadcasting.

  • PDF