• Title/Summary/Keyword: 3D image object recognition

Search Result 128, Processing Time 0.026 seconds

Occluded Object Motion Tracking Method based on Combination of 3D Reconstruction and Optical Flow Estimation (3차원 재구성과 추정된 옵티컬 플로우 기반 가려진 객체 움직임 추적방법)

  • Park, Jun-Heong;Park, Seung-Min;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.5
    • /
    • pp.537-542
    • /
    • 2011
  • A mirror neuron is a neuron fires both when an animal acts and when the animal observes the same action performed by another. We propose a method of 3D reconstruction for occluded object motion tracking like Mirror Neuron System to fire in hidden condition. For modeling system that intention recognition through fire effect like Mirror Neuron System, we calculate depth information using stereo image from a stereo camera and reconstruct three dimension data. Movement direction of object is estimated by optical flow with three-dimensional image data created by three dimension reconstruction. For three dimension reconstruction that enables tracing occluded part, first, picture data was get by stereo camera. Result of optical flow is made be robust to noise by the kalman filter estimation algorithm. Image data is saved as history from reconstructed three dimension image through motion tracking of object. When whole or some part of object is disappeared form stereo camera by other objects, it is restored to bring image date form history of saved past image and track motion of object.

Development of Merging Algorithm between 3-D Objects and Real Image for Augmented Reality

  • Kang, Dong-Joong
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2002.10a
    • /
    • pp.100.5-100
    • /
    • 2002
  • A core technology for implementation of Augmented Reality is to develop a merging algorithm between interesting 3-D objects and real images. In this paper, we present a 3-D object recognition method to decide viewing direction toward the object from camera. This process is the starting point to merge with real image and 3-D objects. Perspective projection between a camera and 3-dimentional objects defines a plane in 3-D space that is from a line in an image and the focal point of the camera. If no errors with perfect 3-D models were introduced in during image feature extraction, then model lines in 3-D space projecting onto this line in the image would exactly lie in this plane. This observa...

  • PDF

Segmentation and Classification of Range Data Using Phase Information of Gabor Fiter (Gabor 필터의 위상 정보를 이용한 거리 영상의 분할 및 분류)

  • 현기호;이광호;황병곤;조석제;하영호
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.27 no.8
    • /
    • pp.1275-1283
    • /
    • 1990
  • Perception of surfaces from range images plays a key role in 3-D object recognition. Recognition of 3-D objects from range images is performed by matching the perceived surface descriptions with stored object models. The first step of the 3-d object recognition from range images is image segmentation. In this paper, an approach for segmenting 3-D range images into symbolic surface descriptions using spatial Gabor filter is proposed. Since the phase of data has a lot of important information, the phase information with magnitude information can effectively segment the range imagery into regions satisfying a common homogeneity criterion. The phase and magnitude of Gabor filter can represent a unique featur vector at a point of range data. As a result, range images are trnasformed into feature vectors in 3-parameter representation. The methods not only to extract meaningful features but also to classify a patch information from range images is presented.

  • PDF

Semantic Object Detection based on LiDAR Distance-based Clustering Techniques for Lightweight Embedded Processors (경량형 임베디드 프로세서를 위한 라이다 거리 기반 클러스터링 기법을 활용한 의미론적 물체 인식)

  • Jung, Dongkyu;Park, Daejin
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.10
    • /
    • pp.1453-1461
    • /
    • 2022
  • The accuracy of peripheral object recognition algorithms using 3D data sensors such as LiDAR in autonomous vehicles has been increasing through many studies, but this requires high performance hardware and complex structures. This object recognition algorithm acts as a large load on the main processor of an autonomous vehicle that requires performing and managing many processors while driving. To reduce this load and simultaneously exploit the advantages of 3D sensor data, we propose 2D data-based recognition using the ROI generated by extracting physical properties from 3D sensor data. In the environment where the brightness value was reduced by 50% in the basic image, it showed 5.3% higher accuracy and 28.57% lower performance time than the existing 2D-based model. Instead of having a 2.46 percent lower accuracy than the 3D-based model in the base image, it has a 6.25 percent reduction in performance time.

Neural Network Approach to Sensor Fusion System for Improving the Recognition Performance of 3D Objects (3차원 물체의 인식 성능 향상을 위한 감각 융합 신경망 시스템)

  • Dong Sung Soo;Lee Chong Ho;Kim Ji Kyoung
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.54 no.3
    • /
    • pp.156-165
    • /
    • 2005
  • Human being recognizes the physical world by integrating a great variety of sensory inputs, the information acquired by their own action, and their knowledge of the world using hierarchically parallel-distributed mechanism. In this paper, authors propose the sensor fusion system that can recognize multiple 3D objects from 2D projection images and tactile informations. The proposed system focuses on improving recognition performance of 3D objects. Unlike the conventional object recognition system that uses image sensor alone, the proposed method uses tactual sensors in addition to visual sensor. Neural network is used to fuse the two sensory signals. Tactual signals are obtained from the reaction force of the pressure sensors at the fingertips when unknown objects are grasped by four-fingered robot hand. The experiment evaluates the recognition rate and the number of learning iterations of various objects. The merits of the proposed systems are not only the high performance of the learning ability but also the reliability of the system with tactual information for recognizing various objects even though the visual sensory signals get defects. The experimental results show that the proposed system can improve recognition rate and reduce teeming time. These results verify the effectiveness of the proposed sensor fusion system as recognition scheme for 3D objects.

Building Information-rich Maps for Intuitive Human Interface Using Networked Knowledge Base

  • Ryu, Jae-Kwan;Kanayama, Chie;Chong, Nak-Young
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1887-1891
    • /
    • 2005
  • Despite significant advances in multimedia transferring technologies in various fields of robotics, it is sometimes quite difficult for the operator to fully understand the context of 3D remote environments from 2D image feedback. Particularly, in the remote control of mobile robots, the recognition of the object associated with the task is very important, because the operator has to control the robot safely in various situations not through trial and error. Therefore, it is necessary to provide the operator with 3D volumetric models of the object and object-related information as well such as locations, shape, size, material properties, and so on. Thus, in this paper, we propose a vision-based human interface system that provides an interactive, information-rich map through network-based information brokering. The system consists of an object recognition part, a 3D map building part, a networked knowledge base part, and a control part of the mobile robot.

  • PDF

Improved Recognition of Far Objects by using DPM method in Curving-Effective Integral Imaging (커브형 집적영상에서 부분적으로 가려진 먼 거리 물체 인식 향상을 위한 DPM 방법)

  • Chung, Han-Gu;Kim, Eun-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.2A
    • /
    • pp.128-134
    • /
    • 2012
  • In this paper, we propose a novel approach to enhance the recognition performance of a far and partially occluded three-dimensional (3-D) target in computational curving-effective integral imaging (CEII) by using the direct pixel-mapping (DPM) method. With this scheme, the elemental image array (EIA) originally picked up from a far and partially occluded 3-D target can be converted into a new EIA just like the one virtually picked up from a target located close to the lenslet array. Due to this characteristic of DPM, resolution and quality of the reconstructed target image can be highly enhanced, which results in a significant improvement of recognition performance of a far 3-D object. Experimental results reveal that image quality of the reconstructed target image and object recognition performance of the proposed system have been improved by 1.75 dB and 4.56% on the average in PSNR (peak-to-peak signal-to-noise ratio) and NCC (normalized correlation coefficient), respectively, compared to the conventional system.

3D Object Recognition Using Appearance Model Space of Feature Point (특징점 Appearance Model Space를 이용한 3차원 물체 인식)

  • Joo, Seong Moon;Lee, Chil Woo
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.2
    • /
    • pp.93-100
    • /
    • 2014
  • 3D object recognition using only 2D images is a difficult work because each images are generated different to according to the view direction of cameras. Because SIFT algorithm defines the local features of the projected images, recognition result is particularly limited in case of input images with strong perspective transformation. In this paper, we propose the object recognition method that improves SIFT algorithm by using several sequential images captured from rotating 3D object around a rotation axis. We use the geometric relationship between adjacent images and merge several images into a generated feature space during recognizing object. To clarify effectiveness of the proposed algorithm, we keep constantly the camera position and illumination conditions. This method can recognize the appearance of 3D objects that previous approach can not recognize with usually SIFT algorithm.

Cooperative recognition using multi-view images

  • Kojoh, Toshiyuki;Nagata, Tadashi;Zha, Hong-Bin
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1993.10b
    • /
    • pp.70-75
    • /
    • 1993
  • We represent a method of 3-D object recognition using multi images in this paper. The recognition process is executed as follows. Object models as prior knowledgement are generated and stored on a computer. To extract features of a recognized object, three CCD cameras are set at vertices of a regular triangle and take images of an object to be recognized. By comparing extracted features with generated models, the object is recognized. In general, it is difficult to recognize 3-D objects because there are the following problems such as how to make the correspondence to both stereo images, generate and store an object model according to a recognition process, and effectively collate information gotten from input images. We resolve these problems using the method that the collation on the basis of features independent on the viewpoint, the generation of object models as enumerating some candidate models in an early recognition level, the execution a tight cooperative process among results gained by analyzing each image. We have made experiments based on real images in which polyhedral objects are used as objects to be recognized. Some of results reveal the usefulness of the proposed method.

  • PDF

DECODE: A Novel Method of DEep CNN-based Object DEtection using Chirps Emission and Echo Signals in Indoor Environment (실내 환경에서 Chirp Emission과 Echo Signal을 이용한 심층신경망 기반 객체 감지 기법)

  • Nam, Hyunsoo;Jeong, Jongpil
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.3
    • /
    • pp.59-66
    • /
    • 2021
  • Humans mainly recognize surrounding objects using visual and auditory information among the five senses (sight, hearing, smell, touch, taste). Major research related to the latest object recognition mainly focuses on analysis using image sensor information. In this paper, after emitting various chirp audio signals into the observation space, collecting echoes through a 2-channel receiving sensor, converting them into spectral images, an object recognition experiment in 3D space was conducted using an image learning algorithm based on deep learning. Through this experiment, the experiment was conducted in a situation where there is noise and echo generated in a general indoor environment, not in the ideal condition of an anechoic room, and the object recognition through echo was able to estimate the position of the object with 83% accuracy. In addition, it was possible to obtain visual information through sound through learning of 3D sound by mapping the inference result to the observation space and the 3D sound spatial signal and outputting it as sound. This means that the use of various echo information along with image information is required for object recognition research, and it is thought that this technology can be used for augmented reality through 3D sound.