• Title/Summary/Keyword: three dimensional vision

Search Result 221, Processing Time 0.028 seconds

Analysis of the Increase of Matching Points for Accuracy Improvement in 3D Reconstruction Using Stereo CCTV Image Data

  • Moon, Kwang-il;Pyeon, MuWook;Eo, YangDam;Kim, JongHwa;Moon, Sujung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.35 no.2
    • /
    • pp.75-80
    • /
    • 2017
  • Recently, there has been growing interest in spatial data that combines information and communication technology with smart cities. The high-precision LiDAR (Light Dectection and Ranging) equipment is mainly used to collect three-dimensional spatial data, and the acquired data is also used to model geographic features and to manage plant construction and cultural heritages which require precision. The LiDAR equipment can collect precise data, but also has limitations because they are expensive and take long time to collect data. On the other hand, in the field of computer vision, research is being conducted on the methods of acquiring image data and performing 3D reconstruction based on image data without expensive equipment. Thus, precise 3D spatial data can be constructed efficiently by collecting and processing image data using CCTVs which are installed as infrastructure facilities in smart cities. However, this method can have an accuracy problem compared to the existing equipment. In this study, experiments were conducted and the results were analyzed to increase the number of extracted matching points by applying the feature-based method and the area-based method in order to improve the precision of 3D spatial data built with image data acquired from stereo CCTVs. For techniques to extract matching points, SIFT algorithm and PATCH algorithm were used. If precise 3D reconstruction is possible using the image data from stereo CCTVs, it will be possible to collect 3D spatial data with low-cost equipment and to collect and build data in real time because image data can be easily acquired through the Web from smart-phones and drones.

Implementation of Stereo Object Tracking Simulator using Optical JTC (광 JTC를 이용한 스테레오 물체추적 시뮬레이터의 구현)

  • Lee, Jae-Soo;Kim, Kyu-Tae;Kim, Eun-Soo
    • Journal of the Korean Institute of Telematics and Electronics D
    • /
    • v.36D no.8
    • /
    • pp.68-78
    • /
    • 1999
  • In the typical stereo vision system, when the focus points of the left and right images are mismatched or the moving object is not in the center of the image, not only the observer can be fatigued & unconscious of three-dimensional effect, but also hard to track the moving object. Therefore, the stereo object tracking system can be used to track the moving object by controlling convergence angle to minimize stereo disparity and controlling pan/tilt to locate moving object in the center of the image. In this paper, as a new approach to stereo object tracking system we introduce a stereo object tracking simulator based on the optical JTC system capable of adaptive tracking. By using this simulator, any kinds of experimental results can be predicted & analyzed and the processing if real-time implementation of stereo object tracking system is suggested through some optical experiments even if background noises exist.

  • PDF

A Method for Reproducing Stereo Images to Adjust Screen Parallax on a 3D Display (3D 디스플레이에서의 화면 시차 제어를 위한 입체 영상재생성 기법)

  • Rhee, Seon-Min;Choi, Jong-Moo;Choi, Soo-Mi
    • Journal of the Korea Computer Graphics Society
    • /
    • v.16 no.4
    • /
    • pp.1-10
    • /
    • 2010
  • We present a method to reproduce in-between views from captured stereo images to control depth feeling that a user can perceive on a 3D display. The stereo images captured from a pair of cameras have a fixed viewpoint and a screen parallax which depend on the physical position and the distance between the cameras. In this paper, we produce stereo images of an intermediate viewpoint between two original cameras by a view interpolation on the input stereo images. Furthermore, the camera separation of the reproduced stereo images can be controlled by a linear interpolation coefficient used by the view interpolation. By using the proposed method, stereo images can be reproduced where the depth feeling and a three dimensional effect is suitable for the individual's eye separation or the characteristic of an application.

Biomimetic approach object detection sensors using multiple imaging (다중 영상을 이용한 생체모방형 물체 접근 감지 센서)

  • Choi, Myoung Hoon;Kim, Min;Jeong, Jae-Hoon;Park, Won-Hyeon;Lee, Dong Heon;Byun, Gi-Sik;Kim, Gwan-Hyung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.05a
    • /
    • pp.91-93
    • /
    • 2016
  • From the 2-D image extracting three-dimensional information as the latter is in the bilateral sibeop using two camera method and when using a monocular camera as a very important step generally as "stereo vision". There in today's CCTV and automatic object tracking system used in many medium much to know the site conditions or work developed more clearly by using a stereo camera that mimics the eyes of humans to maximize the efficiency of avoidance / control start and multiple jobs can do. Object tracking system of the existing 2D image will have but can not recognize the distance to the transition could not be recognized by the observer display using a parallax of a stereo image, and the object can be more effectively controlled.

  • PDF

Robust Hand Region Extraction Using a Joint-based Model (관절 기반의 모델을 활용한 강인한 손 영역 추출)

  • Jang, Seok-Woo;Kim, Sul-Ho;Kim, Gye-Young
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.20 no.9
    • /
    • pp.525-531
    • /
    • 2019
  • Efforts to utilize human gestures to effectively implement a more natural and interactive interface between humans and computers have been ongoing in recent years. In this paper, we propose a new algorithm that accepts consecutive three-dimensional (3D) depth images, defines a hand model, and robustly extracts the human hand region based on six palm joints and 15 finger joints. Then, the 3D depth images are adaptively binarized to exclude non-interest areas, such as the background, and accurately extracts only the hand of the person, which is the area of interest. Experimental results show that the presented algorithm detects only the human hand region 2.4% more accurately than the existing method. The hand region extraction algorithm proposed in this paper is expected to be useful in various practical applications related to computer vision and image processing, such as gesture recognition, virtual reality implementation, 3D motion games, and sign recognition.

Three-Dimensional Convolutional Vision Transformer for Sign Language Translation (수어 번역을 위한 3차원 컨볼루션 비전 트랜스포머)

  • Horyeor Seong;Hyeonjoong Cho
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.3
    • /
    • pp.140-147
    • /
    • 2024
  • In the Republic of Korea, people with hearing impairments are the second-largest demographic within the registered disability community, following those with physical disabilities. Despite this demographic significance, research on sign language translation technology is limited due to several reasons including the limited market size and the lack of adequately annotated datasets. Despite the difficulties, a few researchers continue to improve the performacne of sign language translation technologies by employing the recent advance of deep learning, for example, the transformer architecture, as the transformer-based models have demonstrated noteworthy performance in tasks such as action recognition and video classification. This study focuses on enhancing the recognition performance of sign language translation by combining transformers with 3D-CNN. Through experimental evaluations using the PHOENIX-Wether-2014T dataset [1], we show that the proposed model exhibits comparable performance to existing models in terms of Floating Point Operations Per Second (FLOPs).

Quality Enhancement of 3D Volumetric Contents Based on 6DoF for 5G Telepresence Service

  • Byung-Seo Park;Woosuk Kim;Jin-Kyum Kim;Dong-Wook Kim;Young-Ho Seo
    • Journal of Web Engineering
    • /
    • v.21 no.3
    • /
    • pp.729-750
    • /
    • 2022
  • In general, the importance of 6DoF (degree of freedom) 3D (dimension) volumetric contents technology is emerging in 5G (generation) telepresence service, Web-based (WebGL) graphics, computer vision, robotics, and next-generation augmented reality. Since it is possible to acquire RGB images and depth images in real-time through depth sensors that use various depth acquisition methods such as time of flight (ToF) and lidar, many changes have been made in object detection, tracking, and recognition research. In this paper, we propose a method to improve the quality of 3D models for 5G telepresence by processing images acquired through depth and RGB cameras on a multi-view camera system. In this paper, the quality is improved in two major ways. The first concerns the shape of the 3D model. A method of removing noise outside the object by applying a mask obtained from a color image and a combined filtering operation to obtain the difference in depth information between pixels inside the object were proposed. Second, we propose an illumination compensation method for images acquired through a multi-view camera system for photo-realistic 3D model generation. It is assumed that the three-dimensional volumetric shooting is done indoors, and the location and intensity of illumination according to time are constant. Since the multi-view camera uses a total of 8 pairs and converges toward the center of space, the intensity and angle of light incident on each camera are different even if the illumination is constant. Therefore, all cameras take a color correction chart and use a color optimization function to obtain a color conversion matrix that defines the relationship between the eight acquired images. Using this, the image input from all cameras is corrected based on the color correction chart. It was confirmed that the quality of the 3D model could be improved by effectively removing noise due to the proposed method when acquiring images of a 3D volumetric object using eight cameras. It has been experimentally proven that the color difference between images is reduced.

SAAnnot-C3Pap: Ground Truth Collection Technique of Playing Posture Using Semi Automatic Annotation Method (SAAnnot-C3Pap: 반자동 주석화 방법을 적용한 연주 자세의 그라운드 트루스 수집 기법)

  • Park, So-Hyun;Kim, Seo-Yeon;Park, Young-Ho
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.10
    • /
    • pp.409-418
    • /
    • 2022
  • In this paper, we propose SAAnnot-C3Pap, a semi-automatic annotation method for obtaining ground truth of a player's posture. In order to obtain ground truth about the two-dimensional joint position in the existing music domain, openpose, a two-dimensional posture estimation method, was used or manually labeled. However, automatic annotation methods such as the existing openpose have the disadvantages of showing inaccurate results even though they are fast. Therefore, this paper proposes SAAnnot-C3Pap, a semi-automated annotation method that is a compromise between the two. The proposed approach consists of three main steps: extracting postures using openpose, correcting the parts with errors among the extracted parts using supervisely, and then analyzing the results of openpose and supervisely. Perform the synchronization process. Through the proposed method, it was possible to correct the incorrect 2D joint position detection result that occurred in the openpose, solve the problem of detecting two or more people, and obtain the ground truth in the playing posture. In the experiment, we compare and analyze the results of the semi-automated annotation method openpose and the SAAnnot-C3Pap proposed in this paper. As a result of comparison, the proposed method showed improvement of posture information incorrectly collected through openpose.

Comparison of 3D Space Perception for the Stereoscopic AR Holography (스테레오 증강현실 홀로그래피에서의 삼차원 공간감 비교)

  • Kim, Minju;Wohn, Kwangyun
    • Journal of the HCI Society of Korea
    • /
    • v.8 no.2
    • /
    • pp.21-27
    • /
    • 2013
  • Recently, the use of floating hologram has increased in many different aspects, such as exhibitions, education, advertisements, and so on. Especially, the floating hologram that makes use of half-mirror is widely used. Nevertheless, half-mirror, unfortunately, cannot lead users to the perfect three dimensional hologram experience. Even though it can make the vision look to be up on the air, it does not have the capacity to display itself up on the air, which is the ultimate goal of hologram. In addition, it looks inconsistent when a real object is located behind the half-mirror in order to show the convergence of the two (object and the half-mirror). In this paper, we did the study on comparison of 3D space perception for the stereoscopic AR holography. At first, we applied stereoscopic technology to the half-mirror hologram system for the accurate and realistic AR environment. Then, the users can feel as if the real 3D object behind half-mirror and the reflected virtual image are converged much better in the 3D space. Furthermore, by using depth camera, the location and direction of graphics can be controlled to change depending on the user's point of view. This is the effective way to produce augmented stereoscopic images simply and accurately through half-mirror film without any additional devices. What we saw from the user test were applying 3D images and user interaction leads the users to have 3D spatial awareness and realism more effectively and accurately.

  • PDF

Preliminary Research on the Implementation of Information of Human Facial Part Required for the 3D Printing of Eye Shield (안구차폐체 제작에 필요한 안면부 3차원 정보 구현의 기초연구)

  • Choi, Seokyoon
    • Journal of the Korean Society of Radiology
    • /
    • v.13 no.7
    • /
    • pp.955-960
    • /
    • 2019
  • The Computed tomography (CT) scan can have high radiation in a few tests, and this risk is significant given that it is often repeated in one patient. In children, the incidence of radiation-induced cancer is reported because organs are growing, are more sensitive to radiation. 3D printing has recently been studied to be applied to various applications as a research field for 3D printing applications, research on fabrication of radiation shields and materials has been conducted. The purpose of the 3D printer is to replace the existing panel-type shields and to make customized designs according to the shape of the human body. Therefore, research on 3D information processing to be input to the 3D printer is also necessary. In this study, 3D data of the human body surface, which is the preliminary step of the manufacture of patient-specific eye shield using stereo vision depth map technology, was studied. This study aims to increase the possibility of three-dimensional output. As a result of experimenting with this method, which is relatively simple compared with other methods of 3D information processing, the minimum coordinates for 3D information are extracted. The results of this study provided the advantages and limitations of stereo images using natural light and will be the basic data for the manufacture of eye shields in the future.