• Title/Summary/Keyword: image feature descriptor

Search Result 140, Processing Time 0.027 seconds

Efficient Image Search using Advanced SURF and DCD on Mobile Platform (모바일 플랫폼에서 개선된 SURF와 DCD를 이용한 효율적인 영상 검색)

  • Lee, Yong-Hwan
    • Journal of the Semiconductor & Display Technology
    • /
    • v.14 no.2
    • /
    • pp.53-59
    • /
    • 2015
  • Since the amount of digital image continues to grow in usage, users feel increased difficulty in finding specific images from the image collection. This paper proposes a novel image searching scheme that extracts the image feature using combination of Advanced SURF (Speed-Up Robust Feature) and DCD (Dominant Color Descriptor). The key point of this research is to provide a new feature extraction algorithm to improve the existing SURF method with removal of unnecessary feature in image retrieval, which can be adaptable to mobile system and efficiently run on the mobile environments. To evaluate the proposed scheme, we assessed the performance of simulation in term of average precision and F-score on two databases, commonly used in the field of image retrieval. The experimental results revealed that the proposed algorithm exhibited a significant improvement of over 14.4% in retrieval effectiveness, compared to OpenSURF. The main contribution of this paper is that the proposed approach achieves high accuracy and stability by using ASURF and DCD in searching for natural image on mobile platform.

An Empirical Evaluation of Color Distribution Descriptor for Image Search (이미지 검색을 위한 칼라 분포 기술자의 성능 평가)

  • Lee, Choon-Sang;Lee, Yong-Hwan;Kim, Young-Seop;Rhee, Sang-Burm
    • Journal of the Semiconductor & Display Technology
    • /
    • v.5 no.2 s.15
    • /
    • pp.27-31
    • /
    • 2006
  • As more and more digital images are made by various applications, image retrieval becomes a primary concern in technology of multimedia. This paper presents color based descriptor that uses information of color distribution in color images which is the most basic element for image search and performance of proposed visual feature is evaluated through the simulation. In designing the image search descriptor used color histogram, HSV, Daubechies 9/7 and 2 level wavelet decomposition provide better results than other parameters in terms of computational time and performances. Also histogram quadratic matrix outperforms the sum of absolute difference in similarity measurements, but spends more than 60 computational times.

  • PDF

Enhanced SIFT Descriptor Based on Modified Discrete Gaussian-Hermite Moment

  • Kang, Tae-Koo;Zhang, Huazhen;Kim, Dong W.;Park, Gwi-Tae
    • ETRI Journal
    • /
    • v.34 no.4
    • /
    • pp.572-582
    • /
    • 2012
  • The discrete Gaussian-Hermite moment (DGHM) is a global feature representation method that can be applied to square images. We propose a modified DGHM (MDGHM) method and an MDGHM-based scale-invariant feature transform (MDGHM-SIFT) descriptor. In the MDGHM, we devise a movable mask to represent the local features of a non-square image. The complete set of non-square image features are then represented by the summation of all MDGHMs. We also propose to apply an accumulated MDGHM using multi-order derivatives to obtain distinguishable feature information in the third stage of the SIFT. Finally, we calculate an MDGHM-based magnitude and an MDGHM-based orientation using the accumulated MDGHM. We carry out experiments using the proposed method with six kinds of deformations. The results show that the proposed method can be applied to non-square images without any image truncation and that it significantly outperforms the matching accuracy of other SIFT algorithms.

Feature Matching Algorithm Robust To Viewpoint Change (시점 변화에 강인한 특징점 정합 기법)

  • Jung, Hyun-jo;Yoo, Ji-sang
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.12
    • /
    • pp.2363-2371
    • /
    • 2015
  • In this paper, we propose a new feature matching algorithm which is robust to the viewpoint change by using the FAST(Features from Accelerated Segment Test) feature detector and the SIFT(Scale Invariant Feature Transform) feature descriptor. The original FAST algorithm unnecessarily results in many feature points along the edges in the image. To solve this problem, we apply the principal curvatures for refining it. We use the SIFT descriptor to describe the extracted feature points and calculate the homography matrix through the RANSAC(RANdom SAmple Consensus) with the matching pairs obtained from the two different viewpoint images. To make feature matching robust to the viewpoint change, we classify the matching pairs by calculating the Euclidean distance between the transformed coordinates by the homography transformation with feature points in the reference image and the coordinates of the feature points in the different viewpoint image. Through the experimental results, it is shown that the proposed algorithm has better performance than the conventional feature matching algorithms even though it has much less computational load.

MEGH: A New Affine Invariant Descriptor

  • Dong, Xiaojie;Liu, Erqi;Yang, Jie;Wu, Qiang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.7
    • /
    • pp.1690-1704
    • /
    • 2013
  • An affine invariant descriptor is proposed, which is able to well represent the affine covariant regions. Estimating main orientation is still problematic in many existing method, such as SIFT (scale invariant feature transform) and SURF (speeded up robust features). Instead of aligning the estimated main orientation, in this paper ellipse orientation is directly used. According to ellipse orientation, affine covariant regions are firstly divided into 4 sub-regions with equal angles. Since affine covariant regions are divided from the ellipse orientation, the divided sub-regions are rotation invariant regardless the rotation, if any, of ellipse. Meanwhile, the affine covariant regions are normalized into a circular region. In the end, the gradients of pixels in the circular region are calculated and the partition-based descriptor is created by using the gradients. Compared with the existing descriptors including MROGH, SIFT, GLOH, PCA-SIFT and spin images, the proposed descriptor demonstrates superior performance according to extensive experiments.

Shape Recognition and Classification Based on Poisson Equation- Fourier-Mellin Moment Descriptor

  • Zou, Jian-Cheng;Ke, Nan-Nan;Lu, Yan
    • International Journal of CAD/CAM
    • /
    • v.8 no.1
    • /
    • pp.69-72
    • /
    • 2009
  • In this paper, we present a new shape descriptor, which is named Poisson equation-Fourier-Mellin moment Descriptor. We solve the Poisson equation in the shape area, and use the solution to get feature function, which are then integrated using Fourier-Mellin moment to represent the shape. This method develops the Poisson equation-geometric moment Descriptor proposed by Lena Gorelick, and keeps both advantages of Poisson equation-geometric moment and Fourier-Mellin moment. It is proved better than Poisson equation-geometric moment Descriptor in shape recognition and classification experiments.

MPEG-7 Homogeneous Texture Descriptor

  • Ro, Yong-Man;Kim, Mun-Churl;Kang, Ho-Kyung;Manjunath, B.S.;Kim, Jin-Woong
    • ETRI Journal
    • /
    • v.23 no.2
    • /
    • pp.41-51
    • /
    • 2001
  • MPEG-7 standardization work has started with the aims of providing fundamental tools for describing multimedia contents. MPEG-7 defines the syntax and semantics of descriptors and description schemes so that they may be used as fundamental tools for multimedia content description. In this paper, we introduce a texture based image description and retrieval method, which is adopted as the homogeneous texture descriptor in the visual part of the MPEG-7 final committee draft. The current MPEG-7 homogeneous texture descriptor consists of the mean, the standard deviation value of an image, energy, and energy deviation values of Fourier transform of the image. These are extracted from partitioned frequency channels based on the human visual system (HVS). For reliable extraction of the texture descriptor, Radon transformation is employed. This is suitable for HVS behavior. We also introduce various matching methods; for example, intensity-invariant, rotation-invariant and/or scale-invariant matching. This technique retrieves relevant texture images when the user gives a querying texture image. In order to show the promising performance of the texture descriptor, we take the experimental results with the MPEG-7 test sets. Experimental results show that the MPEG-7 texture descriptor gives an efficient and effective retrieval rate. Furthermore, it gives fast feature extraction time for constructing the texture descriptor.

  • PDF

Image Identifier based on Local Feature's Histogram and Acceleration Technique using GPU (지역 특징 히스토그램 기반 영상식별자와 GPU 가속화)

  • Jeon, Hyeok-June;Seo, Yong-Seok;Hwang, Chi-Jung
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.16 no.9
    • /
    • pp.889-897
    • /
    • 2010
  • Recently, a cutting-edge large-scale image database system has demanded these attributes: search with alarming speed, performs with high accuracy, archives efficiently and much more. An image identifier (descriptor) is for measuring the similarity of two images which plays an important role in this system. The extraction method of an image identifier can be roughly classified into two methods: a local and global method. In this paper, the proposed image identifier, LFH(Local Feature's Histogram), is obtained by a histogram of robust and distinctive local descriptors (features) constrained by a district sub-division of a local region. Furthermore, LFH has not only the properties of a local and global descriptor, but also can perform calculations at a magnificent clip to determine distance with pinpoint accuracy. Additionally, we suggested a way to extract LFH via GPU (OpenGL and GLSL). In this experiment, we have compared the LFH with SIFT (local method) and EHD (global method) via storage capacity, extraction and retrieval time along with accuracy.

Implementation of Object Feature Extraction within Image for Object Tracking (객체 추적을 위한 영상 내의 객체 특징점 추출 알고리즘 구현)

  • Lee, Yong-Hwan;Kim, Youngseop
    • Journal of the Semiconductor & Display Technology
    • /
    • v.17 no.3
    • /
    • pp.113-116
    • /
    • 2018
  • This paper proposes a mobile image search system which uses a sensor information of smart phone, and enables running in a variety of environments, which is implemented on Android platform. The implemented system deals with a new image descriptor using combination of the visual feature (CEDD) with EXIF attributes in the target of JPEG image, and image matching scheme, which is optimized to the mobile platform. Experimental result shows that the proposed method exhibited a significant improved searching results of around 80% in precision in the large image database. Considering the performance such as processing time and precision, we think that the proposed method can be used in other application field.

Study of Feature Based Algorithm Performance Comparison for Image Matching between Virtual Texture Image and Real Image (가상 텍스쳐 영상과 실촬영 영상간 매칭을 위한 특징점 기반 알고리즘 성능 비교 연구)

  • Lee, Yoo Jin;Rhee, Sooahm
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_1
    • /
    • pp.1057-1068
    • /
    • 2022
  • This paper compares the combination performance of feature point-based matching algorithms as a study to confirm the matching possibility between image taken by a user and a virtual texture image with the goal of developing mobile-based real-time image positioning technology. The feature based matching algorithm includes process of extracting features, calculating descriptors, matching features from both images, and finally eliminating mismatched features. At this time, for matching algorithm combination, we combined the process of extracting features and the process of calculating descriptors in the same or different matching algorithm respectively. V-World 3D desktop was used for the virtual indoor texture image. Currently, V-World 3D desktop is reinforced with details such as vertical and horizontal protrusions and dents. In addition, levels with real image textures. Using this, we constructed dataset with virtual indoor texture data as a reference image, and real image shooting at the same location as a target image. After constructing dataset, matching success rate and matching processing time were measured, and based on this, matching algorithm combination was determined for matching real image with virtual image. In this study, based on the characteristics of each matching technique, the matching algorithm was combined and applied to the constructed dataset to confirm the applicability, and performance comparison was also performed when the rotation was additionally considered. As a result of study, it was confirmed that the combination of Scale Invariant Feature Transform (SIFT)'s feature and descriptor detection had the highest matching success rate, but matching processing time was longest. And in the case of Features from Accelerated Segment Test (FAST)'s feature detector and Oriented FAST and Rotated BRIEF (ORB)'s descriptor calculation, the matching success rate was similar to that of SIFT-SIFT combination, while matching processing time was short. Furthermore, in case of FAST-ORB, it was confirmed that the matching performance was superior even when 10° rotation was applied to the dataset. Therefore, it was confirmed that the matching algorithm of FAST-ORB combination could be suitable for matching between virtual texture image and real image.