• Title/Summary/Keyword: Spatial Feature Extraction

Search Result 151, Processing Time 0.025 seconds

A STUDY ON SPATIAL FEATURE EXTRACTION IN THE CLASSIFICATION OF HIGH RESOLUTIION SATELLITE IMAGERY

  • Han, You-Kyung;Kim, Hye-Jin;Choi, Jae-Wan;Kim, Yong-Il
    • Proceedings of the KSRS Conference
    • /
    • 2008.10a
    • /
    • pp.361-364
    • /
    • 2008
  • It is well known that combining spatial and spectral information can improve land use classification from satellite imagery. High spatial resolution classification has a limitation when only using the spectral information due to the complex spatial arrangement of features and spectral heterogeneity within each class. Therefore, extracting the spatial information is one of the most important steps in high resolution satellite image classification. In this paper, we propose a new spatial feature extraction method. The extracted features are integrated with spectral bands to improve overall classification accuracy. The classification is achieved by applying a Support Vector Machines classifier. In order to evaluate the proposed feature extraction method, we applied our approach to KOMPSAT-2 data and compared the result with the other methods.

  • PDF

Development of Digital Surface Model and Feature Extraction by Integrating Laser Scanner and CCD sensor

  • Nagai, Masahiko;Shibasaki, Ryosuke;Zhao, Huijing;Manandhar, Dinesh
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.859-861
    • /
    • 2003
  • In order to present a space in details, it is indispensable to acquire 3D shape and texture simultaneously from the same platform. 3D shape is acquired by Laser Scanner as point cloud data, and texture is acquired by CCD sensor. Positioning data is acquired by IMU (Inertial Measurement Unit). All the sensors and equipments are assembled on a hand-trolley. In this research, a method of integrating the 3D shape and texture for automated construction of Digital Surface Model is developed. This Digital Surface Model is applied for efficient feature extraction. More detailed extraction is possible , because 3D Digital Surface Model has both 3D shape and texture information.

  • PDF

INTERACTIVE FEATURE EXTRACTION FOR IMAGE REGISTRATION

  • Kim Jun-chul;Lee Young-ran;Shin Sung-woong;Kim Kyung-ok
    • Proceedings of the KSRS Conference
    • /
    • 2005.10a
    • /
    • pp.641-644
    • /
    • 2005
  • This paper introduces an Interactive Feature Extraction (!FE) approach for the registration of satellite imagery by matching extracted point and line features. !FE method contains both point extraction by cross-correlation matching of singular points and line extraction by Hough transform. The purpose of this study is to minimize user's intervention in feature extraction and easily apply the extracted features for image registration. Experiments with these imagery dataset proved the feasibility and the efficiency of the suggested method.

  • PDF

Image Description and Matching Scheme Using Synthetic Features for Recommendation Service

  • Yang, Won-Keun;Cho, A-Young;Oh, Weon-Geun;Jeong, Dong-Seok
    • ETRI Journal
    • /
    • v.33 no.4
    • /
    • pp.589-599
    • /
    • 2011
  • This paper presents an image description and matching scheme using synthetic features for a recommendation service. The recommendation service is an example of smart search because it offers something before a user's request. In the proposed extraction scheme, an image is described by synthesized spatial and statistical features. The spatial feature is designed to increase the discriminability by reflecting delicate variations. The statistical feature is designed to increase the robustness by absorbing small variations. For extracting spatial features, we partition the image into concentric circles and extract four characteristics using a spatial relation. To extract statistical features, we adapt three transforms into the image and compose a 3D histogram as the final statistical feature. The matching schemes are designed hierarchically using the proposed spatial and statistical features. The result shows that each feature is better than the compared algorithms that use spatial or statistical features. Additionally, if we adapt the proposed whole extraction and matching scheme, the overall performance will become 98.44% in terms of the correct search ratio.

A Comparison of Global Feature Extraction Technologies and Their Performance for Image Identification (영상 식별을 위한 전역 특징 추출 기술과 그 성능 비교)

  • Yang, Won-Keun;Cho, A-Young;Jeong, Dong-Seok
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.1
    • /
    • pp.1-14
    • /
    • 2011
  • While the circulation of images become active, various requirements to manage increasing database are raised. The content-based technology is one of methods to satisfy these requirements. The image is represented by feature vectors extracted by various methods in the content-based technology. The global feature method insures fast matching speed because the feature vector extracted by the global feature method is formed into a standard shape. The global feature extraction methods are classified into two categories, the spatial feature extraction and statistical feature extraction. And each group is divided by what kind of information is used, color feature or gray scale feature. In this paper, we introduce various global feature extraction technologies and compare their performance by accuracy, recall-precision graph, ANMRR, feature vector size and matching time. According to the experiments, the spatial features show good performance in non-geometrical modifications, and the extraction technologies that use color and histogram feature show the best performance.

A New Covert Visual Attention System by Object-based Spatiotemporal Cues and Their Dynamic Fusioned Saliency Map (객체기반의 시공간 단서와 이들의 동적결합 된돌출맵에 의한 상향식 인공시각주의 시스템)

  • Cheoi, Kyungjoo
    • Journal of Korea Multimedia Society
    • /
    • v.18 no.4
    • /
    • pp.460-472
    • /
    • 2015
  • Most of previous visual attention system finds attention regions based on saliency map which is combined by multiple extracted features. The differences of these systems are in the methods of feature extraction and combination. This paper presents a new system which has an improvement in feature extraction method of color and motion, and in weight decision method of spatial and temporal features. Our system dynamically extracts one color which has the strongest response among two opponent colors, and detects the moving objects not moving pixels. As a combination method of spatial and temporal feature, the proposed system sets the weight dynamically by each features' relative activities. Comparative results show that our suggested feature extraction and integration method improved the detection rate of attention region.

A Study on Optimal Shape-Size Index Extraction for Classification of High Resolution Satellite Imagery (고해상도 영상의 분류결과 개선을 위한 최적의 Shape-Size Index 추출에 관한 연구)

  • Han, You-Kyung;Kim, Hye-Jin;Choi, Jae-Wan;Kim, Yong-Il
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.2
    • /
    • pp.145-154
    • /
    • 2009
  • High spatial resolution satellite image classification has a limitation when only using the spectral information due to the complex spatial arrangement of features and spectral heterogeneity within each class. Therefore, the extraction of the spatial information is one of the most important steps in high resolution satellite image classification. This study proposes a new spatial feature extraction method, named SSI(Shape-Size Index). SSI uses a simple region-growing based image segmentation and allocates spatial property value in each segment. The extracted feature is integrated with spectral bands to improve overall classification accuracy. The classification is achieved by applying a SVM(Support Vector Machines) classifier. In order to evaluate the proposed feature extraction method, KOMPSAT-2 and QuickBird-2 data are used for experiments. It is demonstrated that proposed SSI algorithm leads to a notable increase in classification accuracy.

Semi-fragile Watermarking Scheme for H.264/AVC Video Content Authentication Based on Manifold Feature

  • Ling, Chen;Ur-Rehman, Obaid;Zhang, Wenjun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.12
    • /
    • pp.4568-4587
    • /
    • 2014
  • Authentication of videos and images based on the content is becoming an important problem in information security. Unfortunately, previous studies lack the consideration of Kerckhoffs's principle in order to achieve this (i.e., a cryptosystem should be secure even if everything about the system, except the key, is public knowledge). In this paper, a solution to the problem of finding a relationship between a frame's index and its content is proposed based on the creative utilization of a robust manifold feature. The proposed solution is based on a novel semi-fragile watermarking scheme for H.264/AVC video content authentication. At first, the input I-frame is partitioned for feature extraction and watermark embedding. This is followed by the temporal feature extraction using the Isometric Mapping algorithm. The frame index is included in the feature to produce the temporal watermark. In order to improve security, the spatial watermark will be encrypted together with the temporal watermark. Finally, the resultant watermark is embedded into the Discrete Cosine Transform coefficients in the diagonal positions. At the receiver side, after watermark extraction and decryption, temporal tampering is detected through a mismatch between the frame index extracted from the temporal watermark and the observed frame index. Next, the feature is regenerate through temporal feature regeneration, and compared with the extracted feature. It is judged through the comparison whether the extracted temporal watermark is similar to that of the original watermarked video. Additionally, for spatial authentication, the tampered areas are located via the comparison between extracted and regenerated spatial features. Experimental results show that the proposed method is sensitive to intentional malicious attacks and modifications, whereas it is robust to legitimate manipulations, such as certain level of lossy compression, channel noise, Gaussian filtering and brightness adjustment. Through a comparison between the extracted frame index and the current frame index, the temporal tempering is identified. With the proposed scheme, a solution to the Kerckhoffs's principle problem is specified.

Gesture Extraction for Ubiquitous Robot-Human Interaction (유비쿼터스 로봇과 휴먼 인터액션을 위한 제스쳐 추출)

  • Kim, Moon-Hwan;Joo, Young-Hoon;Park, Jin-Bae
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.11 no.12
    • /
    • pp.1062-1067
    • /
    • 2005
  • This paper discusses a skeleton feature extraction method for ubiquitous robot system. The skeleton features are used to analyze human motion and pose estimation. In different conventional feature extraction environment, the ubiquitous robot system requires more robust feature extraction method because it has internal vibration and low image quality. The new hybrid silhouette extraction method and adaptive skeleton model are proposed to overcome this constrained environment. The skin color is used to extract more sophisticated feature points. Finally, the experimental results show the superiority of the proposed method.

LFFCNN: Multi-focus Image Synthesis in Light Field Camera (LFFCNN: 라이트 필드 카메라의 다중 초점 이미지 합성)

  • Hyeong-Sik Kim;Ga-Bin Nam;Young-Seop Kim
    • Journal of the Semiconductor & Display Technology
    • /
    • v.22 no.3
    • /
    • pp.149-154
    • /
    • 2023
  • This paper presents a novel approach to multi-focus image fusion using light field cameras. The proposed neural network, LFFCNN (Light Field Focus Convolutional Neural Network), is composed of three main modules: feature extraction, feature fusion, and feature reconstruction. Specifically, the feature extraction module incorporates SPP (Spatial Pyramid Pooling) to effectively handle images of various scales. Experimental results demonstrate that the proposed model not only effectively fuses a single All-in-Focus image from images with multi focus images but also offers more efficient and robust focus fusion compared to existing methods.

  • PDF