• Title/Summary/Keyword: SIFT feature

Search Result 231, Processing Time 0.027 seconds

Facial Expression Recognition Using SIFT Descriptor (SIFT 기술자를 이용한 얼굴 표정인식)

  • Kim, Dong-Ju;Lee, Sang-Heon;Sohn, Myoung-Kyu
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.2
    • /
    • pp.89-94
    • /
    • 2016
  • This paper proposed a facial expression recognition approach using SIFT feature and SVM classifier. The SIFT was generally employed as feature descriptor at key-points in object recognition fields. However, this paper applied the SIFT descriptor as feature vector for facial expression recognition. In this paper, the facial feature was extracted by applying SIFT descriptor at each sub-block image without key-point detection procedure, and the facial expression recognition was performed using SVM classifier. The performance evaluation was carried out through comparison with binary pattern feature-based approaches such as LBP and LDP, and the CK facial expression database and the JAFFE facial expression database were used in the experiments. From the experimental results, the proposed method using SIFT descriptor showed performance improvements of 6.06% and 3.87% compared to previous approaches for CK database and JAFFE database, respectively.

Feature Extraction for Endoscopic Image by using the Scale Invariant Feature Transform(SIFT) (SIFT를 이용한 내시경 영상에서의 특징점 추출)

  • Oh, J.S.;Kim, H.C.;Kim, H.R.;Koo, J.M.;Kim, M.G.
    • Proceedings of the KIEE Conference
    • /
    • 2005.10b
    • /
    • pp.6-8
    • /
    • 2005
  • Study that uses geometrical information in computer vision is lively. Problem that should be preceded is matching problem before studying. Feature point should be extracted for well matching. There are a lot of methods that extract feature point from former days are studied. Because problem does not exist algorithm that is applied for all images, it is a hot water. Specially, it is not easy to find feature point in endoscope image. The big problem can not decide easily a point that is predicted feature point as can know even if see endoscope image as eyes. Also, accuracy of matching problem can be decided after number of feature points is enough and also distributed on whole image. In this paper studied algorithm that can apply to endoscope image. SIFT method displayed excellent performance when compared with alternative way (Affine invariant point detector etc.) in general image but SIFT parameter that used in general image can't apply to endoscope image. The gual of this paper is abstraction of feature point on endoscope image that controlled by contrast threshold and curvature threshold among the parameters for applying SIFT method on endoscope image. Studied about method that feature points can have good distribution and control number of feature point than traditional alternative way by controlling the parameters on experiment result.

  • PDF

Comparative Analysis of the Performance of SIFT and SURF (SIFT 와 SURF 알고리즘의 성능적 비교 분석)

  • Lee, Yong-Hwan;Park, Je-Ho;Kim, Youngseop
    • Journal of the Semiconductor & Display Technology
    • /
    • v.12 no.3
    • /
    • pp.59-64
    • /
    • 2013
  • Accurate and robust image registration is important task in many applications such as image retrieval and computer vision. To perform the image registration, essential required steps are needed in the process: feature detection, extraction, matching, and reconstruction of image. In the process of these function, feature extraction not only plays a key role, but also have a big effect on its performance. There are two representative algorithms for extracting image features, which are scale invariant feature transform (SIFT) and speeded up robust feature (SURF). In this paper, we present and evaluate two methods, focusing on comparative analysis of the performance. Experiments for accurate and robust feature detection are shown on various environments such like scale changes, rotation and affine transformation. Experimental trials revealed that SURF algorithm exhibited a significant result in both extracting feature points and matching time, compared to SIFT method.

Effective Marker Placement Method By De Bruijn Sequence for Corresponding Points Matching (드 브루인 수열을 이용한 효과적인 위치 인식 마커 구성)

  • Park, Gyeong-Mi;Kim, Sung-Hwan;Cho, Hwan-Gue
    • The Journal of the Korea Contents Association
    • /
    • v.12 no.6
    • /
    • pp.9-20
    • /
    • 2012
  • In computer vision, it is very important to obtain reliable corresponding feature points. However, we know it is not easy to find the corresponding feature points exactly considering by scaling, lighting, viewpoints, etc. Lots of SIFT methods applies the invariant to image scale and rotation and change in illumination, which is due to the feature vector extracted from corners or edges of object. However, SIFT could not find feature points, if edges do not exist in the area when we extract feature points along edges. In this paper, we present a new placement method of marker to improve the performance of SIFT feature detection and matching between different view of an object or scene. The shape of the markers used in the proposed method is formed in a semicircle to detect dominant direction vector by SIFT algorithm depending on direction placement of marker. We applied De Bruijn sequence for the markers direction placement to improve the matching performance. The experimental results show that the proposed method is more accurate and effective comparing to the current method.

Multiple Vehicle Detection and Tracking in Highway Traffic Surveillance Video Based on SIFT Feature Matching

  • Mu, Kenan;Hui, Fei;Zhao, Xiangmo
    • Journal of Information Processing Systems
    • /
    • v.12 no.2
    • /
    • pp.183-195
    • /
    • 2016
  • This paper presents a complete method for vehicle detection and tracking in a fixed setting based on computer vision. Vehicle detection is performed based on Scale Invariant Feature Transform (SIFT) feature matching. With SIFT feature detection and matching, the geometrical relations between the two images is estimated. Then, the previous image is aligned with the current image so that moving vehicles can be detected by analyzing the difference image of the two aligned images. Vehicle tracking is also performed based on SIFT feature matching. For the decreasing of time consumption and maintaining higher tracking accuracy, the detected candidate vehicle in the current image is matched with the vehicle sample in the tracking sample set, which contains all of the detected vehicles in previous images. Most remarkably, the management of vehicle entries and exits is realized based on SIFT feature matching with an efficient update mechanism of the tracking sample set. This entire method is proposed for highway traffic environment where there are no non-automotive vehicles or pedestrians, as these would interfere with the results.

Integrated SIFT Algorithm with Feature Point Matching Filter for Relative Position Estimation (특징점 정합 필터 결합 SIFT를 이용한 상대 위치 추정)

  • Gwak, Min-Gyu;Sung, Sang-Kyung;Yun, Suk-Chang;Won, Dae-Hee;Lee, Young-Jae
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.37 no.8
    • /
    • pp.759-766
    • /
    • 2009
  • The purpose of this paper is an image processing algorithm development as a base research achieving performance enhancement of integrated navigation system. We used the SIFT (Scale Invariant Feature Transform) algorithm for image processing, and developed feature point matching filter for rejecting mismatched points. By applying the proposed algorithm, it is obtained better result than other methods of parameter tuning and KLT based feature point tracking. For further study, integration with INS and algorithm optimization for the real-time implementation are under investigation.

Constructing 3D Outlines of Objects based on Feature Points using Monocular Camera (단일카메라를 사용한 특징점 기반 물체 3차원 윤곽선 구성)

  • Park, Sang-Heon;Lee, Jeong-Oog;Baik, Doo-Kwon
    • The KIPS Transactions:PartB
    • /
    • v.17B no.6
    • /
    • pp.429-436
    • /
    • 2010
  • This paper presents a method to extract 3D outlines of objects in an image obtained from a monocular vision. After detecting the general outlines of the object by MOPS(Multi-Scale Oriented Patches) -algorithm and we obtain their spatial coordinates. Simultaneously, it obtains the space-coordinates with feature points to be immanent within the outlines of objects through SIFT(Scale Invariant Feature Transform)-algorithm. It grasps a form of objects to join the space-coordinates of outlines and SIFT feature points. The method which is proposed in this paper, it forms general outlines of objects, so that it enables a rapid calculation, and also it has the advantage capable of collecting a detailed data because it supplies the internal-data of outlines through SIFT feature points.

Enhanced SIFT Descriptor Based on Modified Discrete Gaussian-Hermite Moment

  • Kang, Tae-Koo;Zhang, Huazhen;Kim, Dong W.;Park, Gwi-Tae
    • ETRI Journal
    • /
    • v.34 no.4
    • /
    • pp.572-582
    • /
    • 2012
  • The discrete Gaussian-Hermite moment (DGHM) is a global feature representation method that can be applied to square images. We propose a modified DGHM (MDGHM) method and an MDGHM-based scale-invariant feature transform (MDGHM-SIFT) descriptor. In the MDGHM, we devise a movable mask to represent the local features of a non-square image. The complete set of non-square image features are then represented by the summation of all MDGHMs. We also propose to apply an accumulated MDGHM using multi-order derivatives to obtain distinguishable feature information in the third stage of the SIFT. Finally, we calculate an MDGHM-based magnitude and an MDGHM-based orientation using the accumulated MDGHM. We carry out experiments using the proposed method with six kinds of deformations. The results show that the proposed method can be applied to non-square images without any image truncation and that it significantly outperforms the matching accuracy of other SIFT algorithms.

SIFT Image Feature Extraction based on Deep Learning (딥 러닝 기반의 SIFT 이미지 특징 추출)

  • Lee, Jae-Eun;Moon, Won-Jun;Seo, Young-Ho;Kim, Dong-Wook
    • Journal of Broadcast Engineering
    • /
    • v.24 no.2
    • /
    • pp.234-242
    • /
    • 2019
  • In this paper, we propose a deep neural network which extracts SIFT feature points by determining whether the center pixel of a cropped image is a SIFT feature point. The data set of this network consists of a DIV2K dataset cut into $33{\times}33$ size and uses RGB image unlike SIFT which uses black and white image. The ground truth consists of the RobHess SIFT features extracted by setting the octave (scale) to 0, the sigma to 1.6, and the intervals to 3. Based on the VGG-16, we construct an increasingly deep network of 13 to 23 and 33 convolution layers, and experiment with changing the method of increasing the image scale. The result of using the sigmoid function as the activation function of the output layer is compared with the result using the softmax function. Experimental results show that the proposed network not only has more than 99% extraction accuracy but also has high extraction repeatability for distorted images.

Robust AAM-based Face Tracking with Occlusion Using SIFT Features (SIFT 특징을 이용하여 중첩상황에 강인한 AAM 기반 얼굴 추적)

  • Eom, Sung-Eun;Jang, Jun-Su
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.355-362
    • /
    • 2010
  • Face tracking is to estimate the motion of a non-rigid face together with a rigid head in 3D, and plays important roles in higher levels such as face/facial expression/emotion recognition. In this paper, we propose an AAM-based face tracking algorithm. AAM has been widely used to segment and track deformable objects, but there are still many difficulties. Particularly, it often tends to diverge or converge into local minima when a target object is self-occluded, partially or completely occluded. To address this problem, we utilize the scale invariant feature transform (SIFT). SIFT is an effective method for self and partial occlusion because it is able to find correspondence between feature points under partial loss. And it enables an AAM to continue to track without re-initialization in complete occlusions thanks to the good performance of global matching. We also register and use the SIFT features extracted from multi-view face images during tracking to effectively track a face across large pose changes. Our proposed algorithm is validated by comparing other algorithms under the above 3 kinds of occlusions.