• Title/Summary/Keyword: SIFT feature

Search Result 231, Processing Time 0.036 seconds

Deep Learning-based Keypoint Filtering for Remote Sensing Image Registration (원격 탐사 영상 정합을 위한 딥러닝 기반 특징점 필터링)

  • Sung, Jun-Young;Lee, Woo-Ju;Oh, Seoung-Jun
    • Journal of Broadcast Engineering
    • /
    • v.26 no.1
    • /
    • pp.26-38
    • /
    • 2021
  • In this paper, DLKF (Deep Learning Keypoint Filtering), the deep learning-based keypoint filtering method for the rapidization of the image registration method for remote sensing images is proposed. The complexity of the conventional feature-based image registration method arises during the feature matching step. To reduce this complexity, this paper proposes to filter only the keypoints detected in the artificial structure among the keypoints detected in the keypoint detector by ensuring that the feature matching is matched with the keypoints detected in the artificial structure of the image. For reducing the number of keypoints points as preserving essential keypoints, we preserve keypoints adjacent to the boundaries of the artificial structure, and use reduced images, and crop image patches overlapping to eliminate noise from the patch boundary as a result of the image segmentation method. the proposed method improves the speed and accuracy of registration. To verify the performance of DLKF, the speed and accuracy of the conventional keypoints extraction method were compared using the remote sensing image of KOMPSAT-3 satellite. Based on the SIFT-based registration method, which is commonly used in households, the SURF-based registration method, which improved the speed of the SIFT method, improved the speed by 2.6 times while reducing the number of keypoints by about 18%, but the accuracy decreased from 3.42 to 5.43. Became. However, when the proposed method, DLKF, was used, the number of keypoints was reduced by about 82%, improving the speed by about 20.5 times, while reducing the accuracy to 4.51.

Recognition and Modeling of 3D Environment based on Local Invariant Features (지역적 불변특징 기반의 3차원 환경인식 및 모델링)

  • Jang, Dae-Sik
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.3
    • /
    • pp.31-39
    • /
    • 2006
  • This paper presents a novel approach to real-time recognition of 3D environment and objects for various applications such as intelligent robots, intelligent vehicles, intelligent buildings,..etc. First, we establish the three fundamental principles that humans use for recognizing and interacting with the environment. These principles have led to the development of an integrated approach to real-time 3D recognition and modeling, as follows: 1) It starts with a rapid but approximate characterization of the geometric configuration of workspace by identifying global plane features. 2) It quickly recognizes known objects in environment and replaces them by their models in database based on 3D registration. 3) It models the geometric details the geometric details on the fly adaptively to the need of the given task based on a multi-resolution octree representation. SIFT features with their 3D position data, referred to here as stereo-sis SIFT, are used extensively, together with point clouds, for fast extraction of global plane features, for fast recognition of objects, for fast registration of scenes, as well as for overcoming incomplete and noisy nature of point clouds.

  • PDF

Analysis of Computer Vision Application for CGRA Mapping : SIFT (재구성형 프로세서 맵핑을 위한 컴퓨터 비전 응용 분석 : SIFT)

  • Heo, Ingoo;Kim, Yongjoo;Lee, Jinyong;Cho, Yeongpil;Paek, Yunheung;Ko, Kwangman
    • Annual Conference of KIPS
    • /
    • 2011.11a
    • /
    • pp.5-8
    • /
    • 2011
  • 최근 영상이나 이미지로부터 사용자가 원하는 정보를 추출해 내고 재구성 하는 영상 인식, 증강 현실 등의 컴퓨터 비전(Computer Vision) 응용들이 각광을 받고 있다. 이러한 컴퓨터 비전 응용들은 그 동안 많은 알고리즘들의 연구를 통해 꾸준히 개선되고 향상되어 왔으나, 많은 계산량을 요구하기 때문에 임베디드 시스템에서는 널리 쓰이기 힘들었다. 하지만 최근 들어, 스마트폰 등의 모바일 기기에서의 계산 처리 능력이 향상 되고, 소비자 수요가 증가하면서, 이러한 컴퓨터 비전 응용은 점점 모바일 기기에서 널리 쓰이게 되고 있다. 하지만, 여전히 이러한 컴퓨터 응용을 수행하기 위한 계산양은 부족하기 때문에, 충분한 연산량을 제공하기 위한 방법론들이 다양하게 제시되고 있다. 본 논문에서는 이러한 컴퓨터 응용을 위한 프로세서 구조로서 재구성형 프로세서(Reconfigurable Architecture)를 제안한다. 컴퓨터 비전 응용 중 사물 인식 분야에서 널리 쓰이는 SIFT(Scale Invariant Feature Transformation)을 분석하고 이를 재구성형 프로세서에 맵핑하여 성능 향상을 꾀하였다. SIFT의 주요 커널들을 재구성형 프로세서 맵핑한 결과 최소 6.5배에서 최대 9.2배의 성능 향상을 이룰 수 있었다.

A reliable quasi-dense corresponding points for structure from motion

  • Oh, Jangseok;Hong, Hyunggil;Cho, Yongjun;Yun, Haeyong;Seo, Kap-Ho;Kim, Hochul;Kim, Mingi;Lee, Onseok
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3782-3796
    • /
    • 2020
  • A three-dimensional (3D) reconstruction is an important research area in computer vision. The ability to detect and match features across multiple views of a scene is a critical initial step. The tracking matrix W obtained from a 3D reconstruction can be applied to structure from motion (SFM) algorithms for 3D modeling. We often fail to generate an acceptable number of features when processing face or medical images because such images typically contain large homogeneous regions with minimal variation in intensity. In this study, we seek to locate sufficient matching points not only in general images but also in face and medical images, where it is difficult to determine the feature points. The algorithm is implemented on an adaptive threshold value, a scale invariant feature transform (SIFT), affine SIFT, speeded up robust features (SURF), and affine SURF. By applying the algorithm to face and general images and studying the geometric errors, we can achieve quasi-dense matching points that satisfy well-functioning geometric constraints. We also demonstrate a 3D reconstruction with a respectable performance by applying a column space fitting algorithm, which is an SFM algorithm.

Relative Localization for Mobile Robot using 3D Reconstruction of Scale-Invariant Features (스케일불변 특징의 삼차원 재구성을 통한 이동 로봇의 상대위치추정)

  • Kil, Se-Kee;Lee, Jong-Shill;Ryu, Je-Goon;Lee, Eung-Hyuk;Hong, Seung-Hong;Shen, Dong-Fan
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.55 no.4
    • /
    • pp.173-180
    • /
    • 2006
  • A key component of autonomous navigation of intelligent home robot is localization and map building with recognized features from the environment. To validate this, accurate measurement of relative location between robot and features is essential. In this paper, we proposed relative localization algorithm based on 3D reconstruction of scale invariant features of two images which are captured from two parallel cameras. We captured two images from parallel cameras which are attached in front of robot and detect scale invariant features in each image using SIFT(scale invariant feature transform). Then, we performed matching for the two image's feature points and got the relative location using 3D reconstruction for the matched points. Stereo camera needs high precision of two camera's extrinsic and matching pixels in two camera image. Because we used two cameras which are different from stereo camera and scale invariant feature point and it's easy to setup the extrinsic parameter. Furthermore, 3D reconstruction does not need any other sensor. And the results can be simultaneously used by obstacle avoidance, map building and localization. We set 20cm the distance between two camera and capture the 3frames per second. The experimental results show :t6cm maximum error in the range of less than 2m and ${\pm}15cm$ maximum error in the range of between 2m and 4m.

Enhancement on 3 DoF Image Stitching Using Inertia Sensor Data (관성 센서 데이터를 활용한 3 DoF 이미지 스티칭 향상)

  • Kim, Minwoo;Kim, Sang-Kyun
    • Journal of Broadcast Engineering
    • /
    • v.22 no.1
    • /
    • pp.51-61
    • /
    • 2017
  • This paper proposes a method to generate panoramic images by combining conventional feature extraction algorithms (e.g., SIFT, SURF, MPEG-7 CDVS) with sensed data from an inertia sensor to enhance the stitching results. The challenge of image stitching increases when the images are taken from two different mobile phones with no posture calibration. Using inertia sensor data obtained by the mobile phone, images with different yaw angles, pitch angles, roll angles are preprocessed and adjusted before performing stitching process. Performance of stitching (e.g., feature extraction time, inlier point numbers, stitching accuracy) between conventional feature extraction algorithms is reported along with the stitching performance with/without using the inertia sensor data.

Analysis of Shadow Effect on High Resolution Satellite Image Matching in Urban Area (도심지역의 고해상도 위성영상 정합에 대한 그림자 영향 분석)

  • Yeom, Jun Ho;Han, You Kyung;Kim, Yong Il
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.21 no.2
    • /
    • pp.93-98
    • /
    • 2013
  • Multi-temporal high resolution satellite images are essential data for efficient city analysis and monitoring. Yet even when acquired from the same location, identical sensors as well as different sensors, these multi-temporal images have a geometric inconsistency. Matching points between images, therefore, must be extracted to match the images. With images of an urban area, however, it is difficult to extract matching points accurately because buildings, trees, bridges, and other artificial objects cause shadows over a wide area, which have different intensities and directions in multi-temporal images. In this study, we analyze a shadow effect on image matching of high resolution satellite images in urban area using Scale-Invariant Feature Transform(SIFT), the representative matching points extraction method, and automatic shadow extraction method. The shadow segments are extracted using spatial and spectral attributes derived from the image segmentation. Also, we consider information of shadow adjacency with the building edge buffer. SIFT matching points extracted from shadow segments are eliminated from matching point pairs and then image matching is performed. Finally, we evaluate the quality of matching points and image matching results, visually and quantitatively, for the analysis of shadow effect on image matching of high resolution satellite image.

GMM-KL Framework for Indoor Scene Matching (실내 환경 이미지 매칭을 위한 GMM-KL프레임워크)

  • Kim, Jun-Young;Ko, Han-Seok
    • Proceedings of the KIEE Conference
    • /
    • 2005.10b
    • /
    • pp.61-63
    • /
    • 2005
  • Retreiving indoor scene reference image from database using visual information is important issue in Robot Navigation. Scene matching problem in navigation robot is not easy because input image that is taken in navigation process is affinly distorted. We represent probabilistic framework for the feature matching between features in input image and features in database reference images to guarantee robust scene matching efficiency. By reconstructing probabilistic scene matching framework we get a higher precision than the existing feaure-feature matching scheme. To construct probabilistic framework we represent each image as Gaussian Mixture Model using Expectation Maximization algorithm using SIFT(Scale Invariant Feature Transform).

  • PDF

Pre-processing Algorithm for Detection of Slab Information on Steel Process using Robust Feature Points extraction (강건한 특징점 추출을 이용한 철강제품 정보 검출을 위한 전처리 알고리즘)

  • Choi, Jong-Hyun;Yun, Jong-Pil;Choi, Sung-Hoo;Koo, Keun-Hwi;Kim, Sang-Woo
    • Proceedings of the KIEE Conference
    • /
    • 2008.07a
    • /
    • pp.1819-1820
    • /
    • 2008
  • Steel slabs are marked with slab management numbers (SMNs). To increase efficiency, automated identification of SMNs from digital images is desirable. Automatic extraction of SMNs is a prerequisite for automatic character segmentation and recognition. The images include complex background, and the position of the text region of the slabs is variable. This paper describes an pre-processing algorithm for detection of slab information using robust feature points extraction. Using SIFT(Scale Invariant Feature Transform) algorithm, we can reduce the search region for extraction of SMNs from the slab image.

  • PDF

Study on 3 DoF Image and Video Stitching Using Sensed Data

  • Kim, Minwoo;Chun, Jonghoon;Kim, Sang-Kyun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.11 no.9
    • /
    • pp.4527-4548
    • /
    • 2017
  • This paper proposes a method to generate panoramic images by combining conventional feature extraction algorithms (e.g., SIFT, SURF, MPEG-7 CDVS) with sensed data from inertia sensors to enhance the stitching results. The challenge of image stitching increases when the images are taken from two different mobile phones with no posture calibration. Using inertia sensor data obtained by the mobile phone, images with different yaw, pitch, and roll angles are preprocessed and adjusted before performing stitching process. Performance of stitching (e.g., feature extraction time, inlier point numbers, stitching accuracy) between conventional feature extraction algorithms is reported along with the stitching performance with/without using the inertia sensor data. In addition, the stitching accuracy of video data was improved using the same sensed data, with discrete calculation of homograph matrix. The experimental results for stitching accuracies and speed using sensed data are presented in this paper.