• Title/Summary/Keyword: Descriptor Vectors

Search Result 30, Processing Time 0.022 seconds

Multi-Shape Retrieval Using Multi Curvature-Scale Space Descriptor (다중 곡률-단계 공간 기술자를 이용한 다중형상 검색)

  • Park, Sang Hyun;Lee, Soo-Chahn;Yun, Il-Dong
    • Journal of Broadcast Engineering
    • /
    • v.13 no.6
    • /
    • pp.962-965
    • /
    • 2008
  • 2-D shape descriptors, which are vectors representing characteristics of shapes, enable comparison and classification of shapes and are mainly applied to image and 3-D model retrieval. Existing descriptors have limitations that they only describe shapes of single closed contours or lack in precision, making it difficult to be applied to shapes with multiple contours. Therefore, in this paper, we propose a new shape descriptor called Multi-Curvature-Scale Space that can be applied to shapes with multiple contours. Specifically, we represent the topology of the sub-contours in the multi-contour along with Curvature-Scale Space descriptors to represent the shapes of each sub-contours. Also, by allowing the weight of each component to be controlled when computing the distance between descriptors the weight, we deal with ambiguities in measuring similarity between shapes. Results of various experiments that prove the effectiveness of proposed descriptor are presented.

A new approach for content-based video retrieval

  • Kim, Nac-Woo;Lee, Byung-Tak;Koh, Jai-Sang;Song, Ho-Young
    • International Journal of Contents
    • /
    • v.4 no.2
    • /
    • pp.24-28
    • /
    • 2008
  • In this paper, we propose a new approach for content-based video retrieval using non-parametric based motion classification in the shot-based video indexing structure. Our system proposed in this paper has supported the real-time video retrieval using spatio-temporal feature comparison by measuring the similarity between visual features and between motion features, respectively, after extracting representative frame and non-parametric motion information from shot-based video clips segmented by scene change detection method. The extraction of non-parametric based motion features, after the normalized motion vectors are created from an MPEG-compressed stream, is effectively fulfilled by discretizing each normalized motion vector into various angle bins, and by considering the mean, variance, and direction of motion vectors in these bins. To obtain visual feature in representative frame, we use the edge-based spatial descriptor. Experimental results show that our approach is superior to conventional methods with regard to the performance for video indexing and retrieval.

Video Representation via Fusion of Static and Motion Features Applied to Human Activity Recognition

  • Arif, Sheeraz;Wang, Jing;Fei, Zesong;Hussain, Fida
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.7
    • /
    • pp.3599-3619
    • /
    • 2019
  • In human activity recognition system both static and motion information play crucial role for efficient and competitive results. Most of the existing methods are insufficient to extract video features and unable to investigate the level of contribution of both (Static and Motion) components. Our work highlights this problem and proposes Static-Motion fused features descriptor (SMFD), which intelligently leverages both static and motion features in the form of descriptor. First, static features are learned by two-stream 3D convolutional neural network. Second, trajectories are extracted by tracking key points and only those trajectories have been selected which are located in central region of the original video frame in order to to reduce irrelevant background trajectories as well computational complexity. Then, shape and motion descriptors are obtained along with key points by using SIFT flow. Next, cholesky transformation is introduced to fuse static and motion feature vectors to guarantee the equal contribution of all descriptors. Finally, Long Short-Term Memory (LSTM) network is utilized to discover long-term temporal dependencies and final prediction. To confirm the effectiveness of the proposed approach, extensive experiments have been conducted on three well-known datasets i.e. UCF101, HMDB51 and YouTube. Findings shows that the resulting recognition system is on par with state-of-the-art methods.

Content-based image retrieval using region-based image querying (영역 기반의 영상 질의를 이용한 내용 기반 영상 검색)

  • Kim, Nac-Woo;Song, Ho-Young;Kim, Bong-Tae
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.32 no.10C
    • /
    • pp.990-999
    • /
    • 2007
  • In this paper, we propose the region-based image retrieval method using JSEG which is a method for unsupervised segmentation of color-texture regions. JSEG is an algorithm that discretizes an image by color classification, makes the J-image by applying a region to window mask, and then segments the image by using a region growing and merging. The segmented image from JSEG is given to a user as the query image, and a user can select a few segmented regions as the query region. After finding the MBR of regions selected by user query and generating the multiple window masks based on the center point of MBR, we extract the feature vectors from selected regions. We use the accumulated histogram as the global descriptor for performance comparison of extracted feature vectors in each method. Our approach fast and accurately supplies the relevant images for the given query, as the feature vectors extracted from specific regions and global regions are simultaneously applied to image retrieval. Experimental evidence suggests that our algorithm outperforms the recent image-based methods for image indexing and retrieval.

Loitering Detection Solution for CCTV Security System (방범용 CCTV를 위한 배회행위 탐지 솔루션)

  • Kang, Joohyung;Kwak, Sooyeong
    • Journal of Korea Multimedia Society
    • /
    • v.17 no.1
    • /
    • pp.15-25
    • /
    • 2014
  • In this paper, we propose a loitering detection using trajectory probability distribution and local direction descriptor for intelligent surveillance system. We use a background modeling method for detecting moving object and extract the motion features from each moving object for making feature vectors. After that, we detect the loitering behavior person using K-Nearest Neighbor classifier. We test the proposed method in real world environment and it can achieve real time and robust detection results.

Deep Learning-based Scene Change Detection (딥 러닝을 이용한 화면 전환 검출)

  • Lee, Jae-eun;Seo, Young-Ho;Kim, Dong-wook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2019.05a
    • /
    • pp.549-550
    • /
    • 2019
  • In this paper, we propose a method to detect the scene change using deep learning. To extract feature points, we use a deep neural network and express extracted feature points as 128 dimensional vectors using SIFT descriptor. If it is less than 25%, it is determined that the scene is changed.

  • PDF

Medical Image Automatic Annotation Using Multi-class SVM and Annotation Code Array (다중 클래스 SVM과 주석 코드 배열을 이용한 의료 영상 자동 주석 생성)

  • Park, Ki-Hee;Ko, Byoung-Chul;Nam, Jae-Yeal
    • The KIPS Transactions:PartB
    • /
    • v.16B no.4
    • /
    • pp.281-288
    • /
    • 2009
  • This paper proposes a novel algorithm for the efficient classification and annotation of medical images, especially X-ray images. Since X-ray images have a bright foreground against a dark background, we need to extract the different visual descriptors compare with general nature images. In this paper, a Color Structure Descriptor (CSD) based on Harris Corner Detector is only extracted from salient points, and an Edge Histogram Descriptor (EHD) used for a textual feature of image. These two feature vectors are then applied to a multi-class Support Vector Machine (SVM), respectively, to classify images into one of 20 categories. Finally, an image has the Annotation Code Array based on the pre-defined hierarchical relations of categories and priority code order, which is given the several optimal keywords by the Annotation Code Array. Our experiments show that our annotation results have better annotation performance when compared to other method.

Integrating Color, Texture and Edge Features for Content-Based Image Retrieval (내용기반 이미지 검색을 위한 색상, 텍스쳐, 에지 기능의 통합)

  • Ma Ming;Park Dong-Won
    • Science of Emotion and Sensibility
    • /
    • v.7 no.4
    • /
    • pp.57-65
    • /
    • 2004
  • In this paper, we present a hybrid approach which incorporates color, texture and shape in content-based image retrieval. Colors in each image are clustered into a small number of representative colors. The feature descriptor consists of the representative colors and their percentages in the image. A similarity measure similar to the cumulative color histogram distance measure is defined for this descriptor. The co-occurrence matrix as a statistical method is used for texture analysis. An optimal set of five statistical functions are extracted from the co-occurrence matrix of each image, in order to render the feature vector for eachimage maximally informative. The edge information captured within edge histograms is extracted after a pre-processing phase that performs color transformation, quantization, and filtering. The features where thus extracted and stored within feature vectors and were later compared with an intersection-based method. The content-based retrieval system is tested to be effective in terms of retrieval and scalability through experimental results and precision-recall analysis.

  • PDF

Compression Method for MPEG CDVA Global Feature Descriptors (MPEG CDVA 전역 특징 서술자 압축 방법)

  • Kim, Joonsoo;Jo, Won;Lim, Guentaek;Yun, Joungil;Kwak, Sangwoon;Jung, Soon-heung;Cheong, Won-Sik;Choo, Hyon-Gon;Seo, Jeongil;Choi, Yukyung
    • Journal of Broadcast Engineering
    • /
    • v.27 no.3
    • /
    • pp.295-307
    • /
    • 2022
  • In this paper, we propose a novel compression method for scalable Fisher vectors (SCFV) which is used as a global visual feature description of individual video frames in MPEG CDVA standard. CDVA standard has adopted a temporal descriptor redundancy removal technique that takes advantage of the correlation between global feature descriptors for adjacent keyframes. However, due to the variable length property of SCFV, the temporal redundancy removal scheme often results in inferior compression efficiency. It is even worse than the case when the SCFVs are not compressed at all. To enhance the compression efficiency, we propose an asymmetric SCFV difference computation method and a SCFV reconstruction method. Experiments on the FIVR dataset show that the proposed method significantly improves the compression efficiency compared to the original CDVA Experimental Model implementation.

The Weight Decision of Multi-dimensional Features using Fuzzy Similarity Relations and Emotion-Based Music Retrieval (퍼지 유사관계를 이용한 다차원 특징들의 가중치 결정과 감성기반 음악검색)

  • Lim, Jee-Hye;Lee, Joon-Whoan
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.5
    • /
    • pp.637-644
    • /
    • 2011
  • Being digitalized, the music can be easily purchased and delivered to the users. However, there is still some difficulty to find the music which fits to someone's taste using traditional music information search based on musician, genre, tittle, album title and so on. In order to reduce the difficulty, the contents-based or the emotion-based music retrieval has been proposed and developed. In this paper, we propose new method to determine the importance of MPEG-7 low-level audio descriptors which are multi-dimensional vectors for the emotion-based music retrieval. We measured the mutual similarities of musics which represent a pair of emotions expressed by opposite meaning in terms of each multi-dimensional descriptor. Then rough approximation, and inter- and intra similarity ratio from the similarity relation are used for determining the importance of a descriptor, respectively. The set of weights based on the importance decides the aggregated similarity measure, by which emotion-based music retrieval can be achieved. The proposed method shows better result than previous method in terms of the average number of satisfactory musics in the experiment emotion-based retrieval based on content-based search.