• Title/Summary/Keyword: object-based

Search Result 8,051, Processing Time 0.036 seconds

Object Tracking with Sparse Representation based on HOG and LBP Features

  • Boragule, Abhijeet;Yeo, JungYeon;Lee, GueeSang
    • International Journal of Contents
    • /
    • v.11 no.3
    • /
    • pp.47-53
    • /
    • 2015
  • Visual object tracking is a fundamental problem in the field of computer vision, as it needs a proper model to account for drastic appearance changes that are caused by shape, textural, and illumination variations. In this paper, we propose a feature-based visual-object-tracking method with a sparse representation. Generally, most appearance-based models use the gray-scale pixel values of the input image, but this might be insufficient for a description of the target object under a variety of conditions. To obtain the proper information regarding the target object, the following combination of features has been exploited as a corresponding representation: First, the features of the target templates are extracted by using the HOG (histogram of gradient) and LBPs (local binary patterns); secondly, a feature-based sparsity is attained by solving the minimization problems, whereby the target object is represented by the selection of the minimum reconstruction error. The strengths of both features are exploited to enhance the overall performance of the tracker; furthermore, the proposed method is integrated with the particle-filter framework and achieves a promising result in terms of challenging tracking videos.

Multiple Templates and Weighted Correlation Coefficient-based Object Detection and Tracking for Underwater Robots (수중 로봇을 위한 다중 템플릿 및 가중치 상관 계수 기반의 물체 인식 및 추종)

  • Kim, Dong-Hoon;Lee, Dong-Hwa;Myung, Hyun;Choi, Hyun-Taek
    • The Journal of Korea Robotics Society
    • /
    • v.7 no.2
    • /
    • pp.142-149
    • /
    • 2012
  • The camera has limitations of poor visibility in underwater environment due to the limited light source and medium noise of the environment. However, its usefulness in close range has been proved in many studies, especially for navigation. Thus, in this paper, vision-based object detection and tracking techniques using artificial objects for underwater robots have been studied. We employed template matching and mean shift algorithms for the object detection and tracking methods. Also, we propose the weighted correlation coefficient of adaptive threshold -based and color-region-aided approaches to enhance the object detection performance in various illumination conditions. The color information is incorporated into the template matched area and the features of the template are used to robustly calculate correlation coefficients. And the objects are recognized using multi-template matching approach. Finally, the water basin experiments have been conducted to demonstrate the performance of the proposed techniques using an underwater robot platform yShark made by KORDI.

Accelerating particle filter-based object tracking algorithms using parallel programming

  • Truong, Mai Thanh Nhat;Kim, Sanghoon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2018.05a
    • /
    • pp.469-470
    • /
    • 2018
  • Object tracking is a common task in computer vision, an essential part of various vision-based applications. After several years of development, object tracking in video is still a challenging problem because of various visual properties of objects and surrounding environment. Particle filter is a well-known technique among common approaches, has been proven its effectiveness in dealing with difficulties in object tracking. However, particle filter is a high-complexity algorithms, which is an severe disadvantage because object tracking algorithms are required to run in real time. In this research, we utilize parallel programming to accelerate particle filter-based object tracking algorithms. Experimental results showed that our approach reduced the execution time significantly.

Trends on Visual Object Tracking Using Siamese Network (Siamese 네트워크 기반 영상 객체 추적 기술 동향)

  • Oh, J.;Lee, J.
    • Electronics and Telecommunications Trends
    • /
    • v.37 no.1
    • /
    • pp.73-83
    • /
    • 2022
  • Visual object tracking can be utilized in various applications and has attracted considerable attention in the field of computer vision. Visual object tracking technology is classified in various ways based on the number of tracking objects and the methodologies employed for tracking algorithms. This report briefly introduces the visual object tracking challenge that contributes to the development of single object tracking technology. Furthermore, we review ten Siamese network-based algorithms that have attracted attention, owing to their high tracking speed (despite the use of neural networks). In addition, we discuss the prospects of the Siamese network-based object tracking algorithms.

Object Recognition Using 3D RFID System (3D REID 시스템을 이용한 사물 인식)

  • Roh Se-gon;Lee Young Hoon;Choi Hyouk Ryeol
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.11 no.12
    • /
    • pp.1027-1038
    • /
    • 2005
  • Object recognition in the field of robotics generally has depended on a computer vision system. Recently, RFID(Radio Frequency IDentification) has been suggested as technology that supports object recognition. This paper, introduces the advanced RFID-based recognition using a novel tag which is named a 3D tag. The 3D tag was designed to facilitate object recognition. The proposed RFID system not only detects the existence of an object, but also estimates the orientation and position of the object. These characteristics allow the robot to reduce considerably its dependence on other sensors for object recognition. In this paper, we analyze the characteristics of the 3D tag-based RFID system. In addition, the estimation methods of position and orientation using the system are discussed.

Bounding volume estimation algorithm for image-based 3D object reconstruction

  • Jang, Tae Young;Hwang, Sung Soo;Kim, Hee-Dong;Kim, Seong Dae
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.3 no.2
    • /
    • pp.59-64
    • /
    • 2014
  • This paper presents a method for estimating the bounding volume for image-based 3D object reconstruction. The bounding volume of an object is a three-dimensional space where the object is expected to exist, and the size of the bounding volume strongly affects the resolution of the reconstructed geometry. Therefore, the size of a bounding volume should be as small as possible while it encloses an actual object. To this end, the proposed method uses a set of silhouettes of an object and generates a point cloud using a point filter. A bounding volume is then determined as the minimum sphere that encloses the point cloud. The experimental results show that the proposed method generates a bounding volume that encloses an actual object as small as possible.

An Approach to 3D Object Localization Based on Monocular Vision

  • Jung, Sung-Hoon;Jang, Do-Won;Kim, Min-Hwan
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.12
    • /
    • pp.1658-1667
    • /
    • 2008
  • Reconstruction of 3D objects from a single view image is generally an ill-posed problem because of the projection distortion. A monocular vision based 3D object localization method is proposed in this paper, which approximates an object on the ground to a simple bounding solid and works automatically without any prior information about the object. A spherical or cylindrical object determined based on a circularity measure is approximated to a bounding cylinder, while the other general free-shaped objects to a bounding box or a bounding cylinder appropriately. For a general object, its silhouette on the ground is first computed by back-projecting its projected image in image plane onto the ground plane and then a base rectangle on the ground is determined by using the intuition that touched parts of the object on the ground should appear at lower part of the silhouette. The base rectangle is adjusted and extended until a derived bounding box from it can enclose the general object sufficiently. Height of the bounding box is also determined enough to enclose the general object. When the general object looks like a round-shaped object, a bounding cylinder that encloses the bounding box minimally is selected instead of the bounding box. A bounding solid can be utilized to localize a 3D object on the ground and to roughly estimate its volume. Usefulness of our approach is presented with experimental results on real image objects and limitations of our approach are discussed.

  • PDF

A Study on the Object-based Classification Method for Wildfire Fuel Type Map (산불연료지도 제작을 위한 객체기반 분류 방법 연구)

  • Yoon, Yeo-Sang;Kim, Youn-Soo;Kim, Yong-Seung
    • Aerospace Engineering and Technology
    • /
    • v.6 no.1
    • /
    • pp.213-221
    • /
    • 2007
  • This paper showed how to analysis the object-based classification for wildfire fuel type map using Hyperion hyperspectral remote sensing data acquired in April, 2002 and compared the results of the object-based classification with the results of the pixel-based classification. Our methodological approach for wildfire fuel type map firstly processed correcting abnormal pixels and atypical bands and also calibrating atmospheric noise for enhanced image quality. Fuel type map is characterized by the results of the spectral mixture analysis(SMA). Object-based approach was based on segment-based endmember selection, while pixel-based method used standard SMA. To validate and compare, we used true-color high resolution orthoimagery.

  • PDF

Vision-Based Activity Recognition Monitoring Based on Human-Object Interaction at Construction Sites

  • Chae, Yeon;Lee, Hoonyong;Ahn, Changbum R.;Jung, Minhyuk;Park, Moonseo
    • International conference on construction engineering and project management
    • /
    • 2022.06a
    • /
    • pp.877-885
    • /
    • 2022
  • Vision-based activity recognition has been widely attempted at construction sites to estimate productivity and enhance workers' health and safety. Previous studies have focused on extracting an individual worker's postural information from sequential image frames for activity recognition. However, various trades of workers perform different tasks with similar postural patterns, which degrades the performance of activity recognition based on postural information. To this end, this research exploited a concept of human-object interaction, the interaction between a worker and their surrounding objects, considering the fact that trade workers interact with a specific object (e.g., working tools or construction materials) relevant to their trades. This research developed an approach to understand the context from sequential image frames based on four features: posture, object, spatial features, and temporal feature. Both posture and object features were used to analyze the interaction between the worker and the target object, and the other two features were used to detect movements from the entire region of image frames in both temporal and spatial domains. The developed approach used convolutional neural networks (CNN) for feature extractors and activity classifiers and long short-term memory (LSTM) was also used as an activity classifier. The developed approach provided an average accuracy of 85.96% for classifying 12 target construction tasks performed by two trades of workers, which was higher than two benchmark models. This experimental result indicated that integrating a concept of the human-object interaction offers great benefits in activity recognition when various trade workers coexist in a scene.

  • PDF

The Hidden Object Searching Method for Distributed Autonomous Robotic Systems

  • Yoon, Han-Ul;Lee, Dong-Hoon;Sim, Kwee-Bo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1044-1047
    • /
    • 2005
  • In this paper, we present the strategy of object search for distributed autonomous robotic systems (DARS). The DARS are the systems that consist of multiple autonomous robotic agents to whom required functions are distributed. For instance, the agents should recognize their surrounding at where they are located and generate some rules to act upon by themselves. In this paper, we introduce the strategy for multiple DARS robots to search a hidden object at the unknown area. First, we present an area-based action making process to determine the direction change of the robots during their maneuvers. Second, we also present Q learning adaptation to enhance the area-based action making process. Third, we introduce the coordinate system to represent a robot's current location. In the end of this paper, we show experimental results using hexagon-based Q learning to find the hidden object.

  • PDF