• Title/Summary/Keyword: region feature descriptor

Search Result 33, Processing Time 0.021 seconds

Video Representation via Fusion of Static and Motion Features Applied to Human Activity Recognition

  • Arif, Sheeraz;Wang, Jing;Fei, Zesong;Hussain, Fida
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.7
    • /
    • pp.3599-3619
    • /
    • 2019
  • In human activity recognition system both static and motion information play crucial role for efficient and competitive results. Most of the existing methods are insufficient to extract video features and unable to investigate the level of contribution of both (Static and Motion) components. Our work highlights this problem and proposes Static-Motion fused features descriptor (SMFD), which intelligently leverages both static and motion features in the form of descriptor. First, static features are learned by two-stream 3D convolutional neural network. Second, trajectories are extracted by tracking key points and only those trajectories have been selected which are located in central region of the original video frame in order to to reduce irrelevant background trajectories as well computational complexity. Then, shape and motion descriptors are obtained along with key points by using SIFT flow. Next, cholesky transformation is introduced to fuse static and motion feature vectors to guarantee the equal contribution of all descriptors. Finally, Long Short-Term Memory (LSTM) network is utilized to discover long-term temporal dependencies and final prediction. To confirm the effectiveness of the proposed approach, extensive experiments have been conducted on three well-known datasets i.e. UCF101, HMDB51 and YouTube. Findings shows that the resulting recognition system is on par with state-of-the-art methods.

Multiple Pedestrians Tracking using Histogram of Oriented Gradient and Occlusion Detection (기울기 히스토그램 및 폐색 탐지를 통한 다중 보행자 추적)

  • Jeong, Joon-Yong;Jung, Byung-Man;Lee, Kyu-Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.16 no.4
    • /
    • pp.812-820
    • /
    • 2012
  • In this paper, multiple pedestrians tracking system using Histogram of Oriented Gradient and occlusion detection is proposed. The proposed system is applicable to Intelligent Surveillance System. First, we detect pedestrian in a image sequence using pedestrian's feature. To get pedestrian's feature, we make block-histogram using gradient's direction histogram based on HOG(Histogram of Oriented Gradient), after that a pedestrian region is classified by using Linear-SVM(Support Vector Machine) training. Next, moving objects are tracked by using position information of the classified pedestrians. And we create motion trajectory descriptor which is used for content based event retrieval. The experimental results show that the proposed method is more fast, accurate and effective than conventional methods.

Fast Stitching Algorithm by using Feature Tracking (특징점 추적을 통한 다수 영상의 고속 스티칭 기법)

  • Park, Siyoung;Kim, Jongho;Yoo, Jisang
    • Journal of Broadcast Engineering
    • /
    • v.20 no.5
    • /
    • pp.728-737
    • /
    • 2015
  • Stitching algorithm obtain a descriptor of the feature points extracted from multiple images, and create a single image through the matching process between the each of the feature points. In this paper, a feature extraction and matching techniques for the creation of a high-speed panorama using video input is proposed. Features from Accelerated Segment Test(FAST) is used for the feature extraction at high speed. A new feature point matching process, different from the conventional method is proposed. In the matching process, by tracking region containing the feature point through the Mean shift vector required for matching is obtained. Obtained vector is used to match the extracted feature points. In order to remove the outlier, the RANdom Sample Consensus(RANSAC) method is used. By obtaining a homography transformation matrix of the two input images, a single panoramic image is generated. Through experimental results, we show that the proposed algorithm improve of speed panoramic image generation compared to than the existing method.

3D Object Retrieval System Using 2D Shape Information (2차원 모양 정보를 이용한 3차원 물체 검색 시스템)

  • Lim, Sam;Choo, Hyon-Gon;Choi, Min-Seok;Kim, Whoi-Yul
    • Proceedings of the IEEK Conference
    • /
    • 2001.06d
    • /
    • pp.57-60
    • /
    • 2001
  • In this paper, we propose a new 3D object retrieval system using the shape information of 2D silhouette images. 2D images at different view points are derived from a 3D model and linked to the model. Shape feature of 2D image is extracted by a region-based descriptor. In the experiment, we compare the results of the proposed system with those of the system using curvature scale space(CSS) to show the efficiency of our system.

  • PDF

Fast Object Classification Using Texture and Color Information for Video Surveillance Applications (비디오 감시 응용을 위한 텍스쳐와 컬러 정보를 이용한 고속 물체 인식)

  • Islam, Mohammad Khairul;Jahan, Farah;Min, Jae-Hong;Baek, Joong-Hwan
    • Journal of Advanced Navigation Technology
    • /
    • v.15 no.1
    • /
    • pp.140-146
    • /
    • 2011
  • In this paper, we propose a fast object classification method based on texture and color information for video surveillance. We take the advantage of local patches by extracting SURF and color histogram from images. SURF gives intensity content information and color information strengthens distinctiveness by providing links to patch content. We achieve the advantages of fast computation of SURF as well as color cues of objects. We use Bag of Word models to generate global descriptors of a region of interest (ROI) or an image using the local features, and Na$\ddot{i}$ve Bayes model for classifying the global descriptor. In this paper, we also investigate discriminative descriptor named Scale Invariant Feature Transform (SIFT). Our experiment result for 4 classes of the objects shows 95.75% of classification rate.

An Efficient Pedestrian Recognition Method based on PCA Reconstruction and HOG Feature Descriptor (PCA 복원과 HOG 특징 기술자 기반의 효율적인 보행자 인식 방법)

  • Kim, Cheol-Mun;Baek, Yeul-Min;Kim, Whoi-Yul
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.10
    • /
    • pp.162-170
    • /
    • 2013
  • In recent years, the interests and needs of the Pedestrian Protection System (PPS), which is mounted on the vehicle for the purpose of traffic safety improvement is increasing. In this paper, we propose a pedestrian candidate window extraction and unit cell histogram based HOG descriptor calculation methods. At pedestrian detection candidate windows extraction stage, the bright ratio of pedestrian and its circumference region, vertical edge projection, edge factor, and PCA reconstruction image are used. Dalal's HOG requires pixel based histogram calculation by Gaussian weights and trilinear interpolation on overlapping blocks, But our method performs Gaussian down-weight and computes histogram on a per-cell basis, and then the histogram is combined with the adjacent cell, so our method can be calculated faster than Dalal's method. Our PCA reconstruction error based pedestrian detection candidate window extraction method efficiently classifies background based on the difference between pedestrian's head and shoulder area. The proposed method improves detection speed compared to the conventional HOG just using image without any prior information from camera calibration or depth map obtained from stereo cameras.

Shape similarity measure for M:N areal object pairs using the Zernike moment descriptor (저니키 모멘트 서술자를 이용한 M:N 면 객체 쌍의 형상 유사도 측정)

  • Huh, Yong;Yu, Ki-Yun
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.30 no.2
    • /
    • pp.153-162
    • /
    • 2012
  • In this paper, we propose a new shape similarity measure for M:N polygon pairs regardless of different object cardinalities in the pairs. The proposed method compares the projections of two shape functions onto Zernike polynomial basis functions, where the shape functions were obtained from each overall region of objects, thus not being affected by the cardinalities of object pairs. Moments with low-order basis functions describe global shape properties and those with high-order basis functions describe local shape properties. Therefore several moments up to a certain order where the original shapes were similarly reconstructed can efficiently describe the shape properties thus be used for shape comparison. The proposed method was applied for the building objects in the New address digital map and a car navigation map of Seoul area. Comparing to an overlapping ratio method, the proposed method's similarity is more robust to object cardinality.

Face Recognition using Extended Center-Symmetric Pattern and 2D-PCA (Extended Center-Symmetric Pattern과 2D-PCA를 이용한 얼굴인식)

  • Lee, Hyeon Gu;Kim, Dong Ju
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.9 no.2
    • /
    • pp.111-119
    • /
    • 2013
  • Face recognition has recently become one of the most popular research areas in the fields of computer vision, machine learning, and pattern recognition because it spans numerous applications, such as access control, surveillance, security, credit-card verification, and criminal identification. In this paper, we propose a simple descriptor called an ECSP(Extended Center-Symmetric Pattern) for illumination-robust face recognition. The ECSP operator encodes the texture information of a local face region by emphasizing diagonal components of a previous CS-LBP(Center-Symmetric Local Binary Pattern). Here, the diagonal components are emphasized because facial textures along the diagonal direction contain much more information than those of other directions. The facial texture information of the ECSP operator is then used as the input image of an image covariance-based feature extraction algorithm such as 2D-PCA(Two-Dimensional Principal Component Analysis). Performance evaluation of the proposed approach was carried out using various binary pattern operators and recognition algorithms on the Yale B database. The experimental results demonstrated that the proposed approach achieved better recognition accuracy than other approaches, and we confirmed that the proposed approach is effective against illumination variation.

Gabor and Wavelet Texture Descriptors in Representing Textures in Arbitrary Shaped Regions (임의의 영역 안에 텍스처 표현을 위한 Wavelet및 Gabor 텍스처 기술자와 성능평가)

  • Sim Dong-Gyu
    • Journal of Korea Multimedia Society
    • /
    • v.9 no.3
    • /
    • pp.287-295
    • /
    • 2006
  • This paper compares two different approaches based on wavelet and Gabor decomposition towards representing the texture of an arbitrary region. The Gabor-domain mean and standard deviation combination is considered to be best in representing the texture of rectangular regions. However, texture representation of arbitrary regions would enable generalized object-based image retrieval and other applications in the future. In this study, we have found that the wavelet features perform better than the Gabor features in representing the texture of arbitrary regions. Particularly, the wavelet-domain standard deviation and entropy combination results in the best retrieval accuracy. Based on our experiment with texture image sets, we present and compare tile retrieval accuracy of multiple wavelet and Gabor feature combinations.

  • PDF

A Frame-Based Video Signature Method for Very Quick Video Identification and Location

  • Na, Sang-Il;Oh, Weon-Geun;Jeong, Dong-Seok
    • ETRI Journal
    • /
    • v.35 no.2
    • /
    • pp.281-291
    • /
    • 2013
  • A video signature is a set of feature vectors that compactly represents and uniquely characterizes one video clip from another for fast matching. To find a short duplicated region, the video signature must be robust against common video modifications and have a high discriminability. The matching method must be fast and be successful at finding locations. In this paper, a frame-based video signature that uses the spatial information and a two-stage matching method is presented. The proposed method is pair-wise independent and is robust against common video modifications. The proposed two-stage matching method is fast and works very well in finding locations. In addition, the proposed matching structure and strategy can distinguish a case in which a part of the query video matches a part of the target video. The proposed method is verified using video modified by the VCE7 experimental conditions found in MPEG-7. The proposed video signature method achieves a robustness of 88.7% under an independence condition of 5 parts per million with over 1,000 clips being matched per second.