• Title/Summary/Keyword: Tracking feature

Search Result 564, Processing Time 0.035 seconds

Feature point extraction using scale-space filtering and Tracking algorithm based on comparing texturedness similarity (스케일-스페이스 필터링을 통한 특징점 추출 및 질감도 비교를 적용한 추적 알고리즘)

  • Park, Yong-Hee;Kwon, Oh-Seok
    • Journal of Internet Computing and Services
    • /
    • v.6 no.5
    • /
    • pp.85-95
    • /
    • 2005
  • This study proposes a method of feature point extraction using scale-space filtering and a feature point tracking algorithm based on a texturedness similarity comparison, With well-defined operators one can select a scale parameter for feature point extraction; this affects the selection and localization of the feature points and also the performance of the tracking algorithm. This study suggests a feature extraction method using scale-space filtering, With a change in the camera's point of view or movement of an object in sequential images, the window of a feature point will have an affine transform. Traditionally, it is difficult to measure the similarity between correspondence points, and tracking errors often occur. This study also suggests a tracking algorithm that expands Shi-Tomasi-Kanade's tracking algorithm with texturedness similarity.

  • PDF

Face detection using haar-like feature and Tracking with Lucas-Kanade feature tracker (Haar-like feature를 이용한 얼굴 검출과 추적을 위한 Lucas-Kanade특징 추적)

  • Kim, Ki-Sang;Kim, Se-Hoon;Park, Gene-Yong;Choi, Hyung-Il
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.835-838
    • /
    • 2008
  • In this paper, we present automatic face detection and tracking which is robustness in rotation and translation. Detecting a face image, we used Haar-like feature, which is fast detect facial image. Also tracking, we applied Lucas-Kanade feature tracker and KLT algorithm, which has robustness for rotated facial image. In experiment result, we confirmed that face detection and tracking which is robustness in rotation and translation.

  • PDF

Robust Feature Extraction and Tracking Algorithm Using 2-dimensional Wavelet Transform (2차원 웨이브릿 변환을 이용한 강건한 특징점 추출 및 추적 알고리즘)

  • Jang, Sung-Kun;Suk, Jung-Youp
    • Proceedings of the IEEK Conference
    • /
    • 2007.07a
    • /
    • pp.405-406
    • /
    • 2007
  • In this paper, we propose feature extraction and tracking algorithm using multi resolution in 2-dimensional wavelet domain. Feature extraction selects feature points using 2-level wavelet transform in interested region. Feature tracking estimates displacement between current frame and next frame based on feature point which is selected feature extraction algorithm. Experimental results show that the proposed algorithm confirmed a better performance than the existing other algorithms.

  • PDF

Multiple Vehicle Detection and Tracking in Highway Traffic Surveillance Video Based on SIFT Feature Matching

  • Mu, Kenan;Hui, Fei;Zhao, Xiangmo
    • Journal of Information Processing Systems
    • /
    • v.12 no.2
    • /
    • pp.183-195
    • /
    • 2016
  • This paper presents a complete method for vehicle detection and tracking in a fixed setting based on computer vision. Vehicle detection is performed based on Scale Invariant Feature Transform (SIFT) feature matching. With SIFT feature detection and matching, the geometrical relations between the two images is estimated. Then, the previous image is aligned with the current image so that moving vehicles can be detected by analyzing the difference image of the two aligned images. Vehicle tracking is also performed based on SIFT feature matching. For the decreasing of time consumption and maintaining higher tracking accuracy, the detected candidate vehicle in the current image is matched with the vehicle sample in the tracking sample set, which contains all of the detected vehicles in previous images. Most remarkably, the management of vehicle entries and exits is realized based on SIFT feature matching with an efficient update mechanism of the tracking sample set. This entire method is proposed for highway traffic environment where there are no non-automotive vehicles or pedestrians, as these would interfere with the results.

Fast Natural Feature Tracking Using Optical Flow (광류를 사용한 빠른 자연특징 추적)

  • Bae, Byung-Jo;Park, Jong-Seung
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.345-354
    • /
    • 2010
  • Visual tracking techniques for Augmented Reality are classified as either a marker tracking approach or a natural feature tracking approach. Marker-based tracking algorithms can be efficiently implemented sufficient to work in real-time on mobile devices. On the other hand, natural feature tracking methods require a lot of computationally expensive procedures. Most previous natural feature tracking methods include heavy feature extraction and pattern matching procedures for each of the input image frame. It is difficult to implement real-time augmented reality applications including the capability of natural feature tracking on low performance devices. The required computational time cost is also in proportion to the number of patterns to be matched. To speed up the natural feature tracking process, we propose a novel fast tracking method based on optical flow. We implemented the proposed method on mobile devices to run in real-time and be appropriately used with mobile augmented reality applications. Moreover, during tracking, we keep up the total number of feature points by inserting new feature points proportional to the number of vanished feature points. Experimental results showed that the proposed method reduces the computational cost and also stabilizes the camera pose estimation results.

Robust 2D Feature Tracking in Long Video Sequences (긴 비디오 프레임들에서의 강건한 2차원 특징점 추적)

  • Yoon, Jong-Hyun;Park, Jong-Seung
    • The KIPS Transactions:PartB
    • /
    • v.14B no.7
    • /
    • pp.473-480
    • /
    • 2007
  • Feature tracking in video frame sequences has suffered from the instability and the frequent failure of feature matching between two successive frames. In this paper, we propose a robust 2D feature tracking method that is stable to long video sequences. To improve the stability of feature tracking, we predict the spatial movement in the current image frame using the state variables. The predicted current movement is used for the initialization of the search window. By computing the feature similarities in the search window, we refine the current feature positions. Then, the current feature states are updated. This tracking process is repeated for each input frame. To reduce false matches, the outlier rejection stage is also introduced. Experimental results from real video sequences showed that the proposed method performs stable feature tracking for long frame sequences.

Robust appearance feature learning using pixel-wise discrimination for visual tracking

  • Kim, Minji;Kim, Sungchan
    • ETRI Journal
    • /
    • v.41 no.4
    • /
    • pp.483-493
    • /
    • 2019
  • Considering the high dimensions of video sequences, it is often challenging to acquire a sufficient dataset to train the tracking models. From this perspective, we propose to revisit the idea of hand-crafted feature learning to avoid such a requirement from a dataset. The proposed tracking approach is composed of two phases, detection and tracking, according to how severely the appearance of a target changes. The detection phase addresses severe and rapid variations by learning a new appearance model that classifies the pixels into foreground (or target) and background. We further combine the raw pixel features of the color intensity and spatial location with convolutional feature activations for robust target representation. The tracking phase tracks a target by searching for frame regions where the best pixel-level agreement to the model learned from the detection phase is achieved. Our two-phase approach results in efficient and accurate tracking, outperforming recent methods in various challenging cases of target appearance changes.

Object Tracking using Feature Map from Convolutional Neural Network (컨볼루션 신경망의 특징맵을 사용한 객체 추적)

  • Lim, Suchang;Kim, Do Yeon
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.2
    • /
    • pp.126-133
    • /
    • 2017
  • The conventional hand-crafted features used to track objects have limitations in object representation. Convolutional neural networks, which show good performance results in various areas of computer vision, are emerging as new ways to break through the limitations of feature extraction. CNN extracts the features of the image through layers of multiple layers, and learns the kernel used for feature extraction by itself. In this paper, we use the feature map extracted from the convolution layer of the convolution neural network to create an outline model of the object and use it for tracking. We propose a method to adaptively update the outline model to cope with various environment change factors affecting the tracking performance. The proposed algorithm evaluated the validity test based on the 11 environmental change attributes of the CVPR2013 tracking benchmark and showed excellent results in six attributes.

Feature tracking algorithm using multi resolution in wavelet transform domain (웨이브릿 변환 영역에서 다중 해상도를 이용한 특징점 추적 알고리즘)

  • Jang, Sung-Kun;Suk, Jung-Youp;Jin, Sang-Hun;Kim, Sung-Un;Yeo, Bo-Yeon
    • Proceedings of the IEEK Conference
    • /
    • 2006.06a
    • /
    • pp.447-448
    • /
    • 2006
  • In this paper, we propose tracking algorithm using multi resolution in wavelet transform domain. This algorithm consists of two steps. The first step is feature extraction that is select feature-points using 1-level wavelet transform in ROI (Region of Interest). The other step is feature tracking. Based on multi resolution of wavelet transform, we estimate a displacement between current frame and next frame on the basis of selected feature-points. Experimental results show that the proposed algorithm confirmed a better performance than a centroid tracking and correlation tracking.

  • PDF

Non-Prior Training Active Feature Model-Based Object Tracking for Real-Time Surveillance Systems (실시간 감시 시스템을 위한 사전 무학습 능동 특징점 모델 기반 객체 추적)

  • 김상진;신정호;이성원;백준기
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.41 no.5
    • /
    • pp.23-34
    • /
    • 2004
  • In this paper we propose a feature point tracking algorithm using optical flow under non-prior taming active feature model (NPT-AFM). The proposed algorithm mainly focuses on analysis non-rigid objects[1], and provides real-time, robust tracking by NPT-AFM. NPT-AFM algorithm can be divided into two steps: (i) localization of an object-of-interest and (ii) prediction and correction of the object position by utilizing the inter-frame information. The localization step was realized by using a modified Shi-Tomasi's feature tracking algoriam[2] after motion-based segmentation. In the prediction-correction step, given feature points are continuously tracked by using optical flow method[3] and if a feature point cannot be properly tracked, temporal and spatial prediction schemes can be employed for that point until it becomes uncovered again. Feature points inside an object are estimated instead of its shape boundary, and are updated an element of the training set for AFH Experimental results, show that the proposed NPT-AFM-based algerian can robustly track non-rigid objects in real-time.