• Title/Summary/Keyword: 3D Object Recognition

Search Result 272, Processing Time 0.031 seconds

Dual Autostereoscopic Display Platform for Multi-user Collaboration with Natural Interaction

  • Kim, Hye-Mi;Lee, Gun-A.;Yang, Ung-Yeon;Kwak, Tae-Jin;Kim, Ki-Hong
    • ETRI Journal
    • /
    • v.34 no.3
    • /
    • pp.466-469
    • /
    • 2012
  • In this letter, we propose a dual autostereoscopic display platform employing a natural interaction method, which will be useful for sharing visual data with users. To provide 3D visualization of a model to users who collaborate with each other, a beamsplitter is used with a pair of autostereoscopic displays, providing a visual illusion of a floating 3D image. To interact with the virtual object, we track the user's hands with a depth camera. The gesture recognition technique we use operates without any initialization process, such as specific poses or gestures, and supports several commands to control virtual objects by gesture recognition. Experiment results show that our system performs well in visualizing 3D models in real-time and handling them under unconstrained conditions, such as complicated backgrounds or a user wearing short sleeves.

Spherical Point Tracing for Synthetic Vehicle Data Generation with 3D LiDAR Point Cloud Data (3차원 LiDAR 점군 데이터에서의 가상 차량 데이터 생성을 위한 구면 점 추적 기법)

  • Sangjun Lee;Hakil Kim
    • Journal of Broadcast Engineering
    • /
    • v.28 no.3
    • /
    • pp.329-332
    • /
    • 2023
  • 3D Object Detection using deep neural network has been developed a lot for obstacle detection in autonomous vehicles because it can recognize not only the class of target object but also the distance from the object. But in the case of 3D Object Detection models, the detection performance for distant objects is lower than that for nearby objects, which is a critical issue for autonomous vehicles. In this paper, we introduce a technique that increases the performance of 3D object detection models, particularly in recognizing distant objects, by generating virtual 3D vehicle data and adding it to the dataset used for model training. We used a spherical point tracing method that leverages the characteristics of 3D LiDAR sensor data to create virtual vehicles that closely resemble real ones, and we demonstrated the validity of the virtual data by using it to improve recognition performance for objects at all distances in model training.

Predicting Unseen Object Pose with an Adaptive Depth Estimator (적응형 깊이 추정기를 이용한 미지 물체의 자세 예측)

  • Sungho, Song;Incheol, Kim
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.11 no.12
    • /
    • pp.509-516
    • /
    • 2022
  • Accurate pose prediction of objects in 3D space is an important visual recognition technique widely used in many applications such as scene understanding in both indoor and outdoor environments, robotic object manipulation, autonomous driving, and augmented reality. Most previous works for object pose estimation have the limitation that they require an exact 3D CAD model for each object. Unlike such previous works, this paper proposes a novel neural network model that can predict the poses of unknown objects based on only their RGB color images without the corresponding 3D CAD models. The proposed model can obtain depth maps required for unknown object pose prediction by using an adaptive depth estimator, AdaBins,. In this paper, we evaluate the usefulness and the performance of the proposed model through experiments using benchmark datasets.

Optical Scanning Holography - A Review of Recent Progress

  • Poon, Ting-Chung
    • Journal of the Optical Society of Korea
    • /
    • v.13 no.4
    • /
    • pp.406-415
    • /
    • 2009
  • Optical scanning holography (OSH) is a distinct digital holographic technique in that real-time holographic recording a three-dimensional (3-D) object can be acquired by using two-dimensional active optical heterodyne scanning. Applications of the technique so far have included optical scanning cryptography, optical scanning microscopy, 3-D pattern recognition, 3-D holographic TV, and 3-D optical remote sensing. This paper reviews some of the recent progress in OSH. Some possible further works are also discussed.

Visual Positioning System based on Voxel Labeling using Object Simultaneous Localization And Mapping

  • Jung, Tae-Won;Kim, In-Seon;Jung, Kye-Dong
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.302-306
    • /
    • 2021
  • Indoor localization is one of the basic elements of Location-Based Service, such as indoor navigation, location-based precision marketing, spatial recognition of robotics, augmented reality, and mixed reality. We propose a Voxel Labeling-based visual positioning system using object simultaneous localization and mapping (SLAM). Our method is a method of determining a location through single image 3D cuboid object detection and object SLAM for indoor navigation, then mapping to create an indoor map, addressing it with voxels, and matching with a defined space. First, high-quality cuboids are created from sampling 2D bounding boxes and vanishing points for single image object detection. And after jointly optimizing the poses of cameras, objects, and points, it is a Visual Positioning System (VPS) through matching with the pose information of the object in the voxel database. Our method provided the spatial information needed to the user with improved location accuracy and direction estimation.

Sell-modeling of Cylindrical Object based on Generic Model for 3D Object Recognition (3 차원 물체 인식을 위한 보편적 지식기반 실린더형 물체 자가모델링 기법)

  • Baek, Kyeong-Keun;Park, Yeon-Chool;Park, Joon-Young;Lee, Suk-Han
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.210-214
    • /
    • 2008
  • It is actually impossible to model and store all objects which exist in real home environment into robot's database in advance. To resolve this problem, this paper proposes new object modeling method that can be available for robot self-modeling, which is capable of estimating whole model's shape from partial surface data using Generic Model. And this whole produce is conducted to cylindrical objects like cup, bottles and cans which can be easily found at indoor environment. The detailed process is firstly we obtain cylinder's initial principle axis using points coordinates and normal vectors from object's surface after we separate cylindrical object from 3D image. This 3D image is obtained from 3D sensor. And second, we compensate errors in the principle axis repeatedly. Then finally, we do modeling whole cylindrical object using cross sectional principal axis and its radius To show the feasibility of the algorithm, We implemented it and evaluated its accuracy.

  • PDF

Development of a Real-Time 3D Object Detection System using a Deep Learning-based 2D Object Recognition Model and Low-Cost LiDAR Sensor (딥러닝 기반 2D 객체 인식 모델과 저비용 LiDAR 센서를 이용한 실시간 3D 객체 탐지 시스템 개발)

  • Aejin Lee;Yejin Hwang;Boin Jeong;Ki Yong Lee
    • Annual Conference of KIPS
    • /
    • 2023.11a
    • /
    • pp.716-717
    • /
    • 2023
  • 최근 자율주행 기술이 큰 주목을 받고 있지만 고가의 센서를 필요로 하기 때문에 연구 및 상용화에 큰 어려움을 겪고 있다. 따라서 본 논문은 쉽게 사용 가능한 딥러닝 2D 객체 인식 모델과 범용 태블릿에 탑재된 저비용 LiDAR 센서를 이용하여 실시간 3D 객체 탐지가 가능한 시스템을 개발한다. 개발된 시스템을 실제 1/10 크기의 차량 모델에 적용하여 테스트해본 결과 개발 용이성과 정확도 측면에서 자율주행을 위한 저비용 센서로 충분히 활용될 가능성이 있음을 확인하였다.

Development of a Robot arm capable of recognizing 3-D object using stereo vision

  • Kim, Sungjin;Park, Seungjun;Park, Hongphyo;Sangchul Won
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2001.10a
    • /
    • pp.128.6-128
    • /
    • 2001
  • In this paper, we present a methodology of sensing and control for a robot system designed to be capable of grasping an object and moving it to target point Stereo vision system is employed to determine to depth map which represents the distance from the camera. In stereo vision system we have used a center-referenced projection to represent the discrete match space for stereo correspondence. This center-referenced disparity space contains new occlusion points in addition to the match points which we exploit to create a concise representation of correspondence an occlusion. And from the depth map we find the target object´s pose and position in 3-D space. To find the target object´s pose and position, we use the method of the model-based recognition.

  • PDF

Object Detection Based on Hellinger Distance IoU and Objectron Application (Hellinger 거리 IoU와 Objectron 적용을 기반으로 하는 객체 감지)

  • Kim, Yong-Gil;Moon, Kyung-Il
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.2
    • /
    • pp.63-70
    • /
    • 2022
  • Although 2D Object detection has been largely improved in the past years with the advance of deep learning methods and the use of large labeled image datasets, 3D object detection from 2D imagery is a challenging problem in a variety of applications such as robotics, due to the lack of data and diversity of appearances and shapes of objects within a category. Google has just announced the launch of Objectron that has a novel data pipeline using mobile augmented reality session data. However, it also is corresponding to 2D-driven 3D object detection technique. This study explores more mature 2D object detection method, and applies its 2D projection to Objectron 3D lifting system. Most object detection methods use bounding boxes to encode and represent the object shape and location. In this work, we explore a stochastic representation of object regions using Gaussian distributions. We also present a similarity measure for the Gaussian distributions based on the Hellinger Distance, which can be viewed as a stochastic Intersection-over-Union. Our experimental results show that the proposed Gaussian representations are closer to annotated segmentation masks in available datasets. Thus, less accuracy problem that is one of several limitations of Objectron can be relaxed.

Generating 3D Digital Twins of Real Indoor Spaces based on Real-World Point Cloud Data

  • Wonseop Shin;Jaeseok Yoo;Bumsoo Kim;Yonghoon Jung;Muhammad Sajjad;Youngsup Park;Sanghyun Seo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.8
    • /
    • pp.2381-2398
    • /
    • 2024
  • The construction of virtual indoor spaces is crucial for the development of metaverses, virtual production, and other 3D content domains. Traditional methods for creating these spaces are often cost-prohibitive and labor-intensive. To address these challenges, we present a pipeline for generating digital twins of real indoor environments from RGB-D camera-scanned data. Our pipeline synergizes space structure estimation, 3D object detection, and the inpainting of missing areas, utilizing deep learning technologies to automate the creation process. Specifically, we apply deep learning models for object recognition and area inpainting, significantly enhancing the accuracy and efficiency of virtual space construction. Our approach minimizes manual labor and reduces costs, paving the way for the creation of metaverse spaces that closely mimic real-world environments. Experimental results demonstrate the effectiveness of our deep learning applications in overcoming traditional obstacles in digital twin creation, offering high-fidelity digital replicas of indoor spaces. This advancement opens for immersive and realistic virtual content creation, showcasing the potential of deep learning in the field of virtual space construction.