• Title/Summary/Keyword: pixel based classification

Search Result 173, Processing Time 0.033 seconds

Object Classification Method using Hilbert Scanning Distance (힐버트 스캔 거리값을 이용한 물체식별 알고리즘)

  • Choi, Jeong-Hwan;Baek, Young-Min;Choi, Jin-Young
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.57 no.4
    • /
    • pp.700-705
    • /
    • 2008
  • In this paper, we propose object classification algorithm for real-time surveillance system. We have approached this problem using silhouette-based template matching. The silhouette of the object is extracted, and then it is compared with representative template models. Template models are previously stored in the database. Our algorithm is similar to previous pixel-based template matching scheme like Hausdorff Distance, but we use 1D image array rather than 2D regions inspired by Hilbert Path. Transformation of images could reduce computational burden to compute similarity between the detected image and the template images. Experimental results show robustness and real-time performance in object classification, even in low resolution images.

Application of Multispectral Remotely Sensed Imagery for the Characterization of Complex Coastal Wetland Ecosystems of southern India: A Special Emphasis on Comparing Soft and Hard Classification Methods

  • Shanmugam, Palanisamy;Ahn, Yu-Hwan;Sanjeevi , Shanmugam
    • Korean Journal of Remote Sensing
    • /
    • v.21 no.3
    • /
    • pp.189-211
    • /
    • 2005
  • This paper makes an effort to compare the recently evolved soft classification method based on Linear Spectral Mixture Modeling (LSMM) with the traditional hard classification methods based on Iterative Self-Organizing Data Analysis (ISODATA) and Maximum Likelihood Classification (MLC) algorithms in order to achieve appropriate results for mapping, monitoring and preserving valuable coastal wetland ecosystems of southern India using Indian Remote Sensing Satellite (IRS) 1C/1D LISS-III and Landsat-5 Thematic Mapper image data. ISODATA and MLC methods were attempted on these satellite image data to produce maps of 5, 10, 15 and 20 wetland classes for each of three contrast coastal wetland sites, Pitchavaram, Vedaranniyam and Rameswaram. The accuracy of the derived classes was assessed with the simplest descriptive statistic technique called overall accuracy and a discrete multivariate technique called KAPPA accuracy. ISODATA classification resulted in maps with poor accuracy compared to MLC classification that produced maps with improved accuracy. However, there was a systematic decrease in overall accuracy and KAPPA accuracy, when more number of classes was derived from IRS-1C/1D and Landsat-5 TM imagery by ISODATA and MLC. There were two principal factors for the decreased classification accuracy, namely spectral overlapping/confusion and inadequate spatial resolution of the sensors. Compared to the former, the limited instantaneous field of view (IFOV) of these sensors caused occurrence of number of mixture pixels (mixels) in the image and its effect on the classification process was a major problem to deriving accurate wetland cover types, in spite of the increasing spatial resolution of new generation Earth Observation Sensors (EOS). In order to improve the classification accuracy, a soft classification method based on Linear Spectral Mixture Modeling (LSMM) was described to calculate the spectral mixture and classify IRS-1C/1D LISS-III and Landsat-5 TM Imagery. This method considered number of reflectance end-members that form the scene spectra, followed by the determination of their nature and finally the decomposition of the spectra into their endmembers. To evaluate the LSMM areal estimates, resulted fractional end-members were compared with normalized difference vegetation index (NDVI), ground truth data, as well as those estimates derived from the traditional hard classifier (MLC). The findings revealed that NDVI values and vegetation fractions were positively correlated ($r^2$= 0.96, 0.95 and 0.92 for Rameswaram, Vedaranniyam and Pitchavaram respectively) and NDVI and soil fraction values were negatively correlated ($r^2$ =0.53, 0.39 and 0.13), indicating the reliability of the sub-pixel classification. Comparing with ground truth data, the precision of LSMM for deriving moisture fraction was 92% and 96% for soil fraction. The LSMM in general would seem well suited to locating small wetland habitats which occurred as sub-pixel inclusions, and to representing continuous gradations between different habitat types.

Classification of Textured Images Based on Discrete Wavelet Transform and Information Fusion

  • Anibou, Chaimae;Saidi, Mohammed Nabil;Aboutajdine, Driss
    • Journal of Information Processing Systems
    • /
    • v.11 no.3
    • /
    • pp.421-437
    • /
    • 2015
  • This paper aims to present a supervised classification algorithm based on data fusion for the segmentation of the textured images. The feature extraction method we used is based on discrete wavelet transform (DWT). In the segmentation stage, the estimated feature vector of each pixel is sent to the support vector machine (SVM) classifier for initial labeling. To obtain a more accurate segmentation result, two strategies based on information fusion were used. We first integrated decision-level fusion strategies by combining decisions made by the SVM classifier within a sliding window. In the second strategy, the fuzzy set theory and rules based on probability theory were used to combine the scores obtained by SVM over a sliding window. Finally, the performance of the proposed segmentation algorithm was demonstrated on a variety of synthetic and real images and showed that the proposed data fusion method improved the classification accuracy compared to applying a SVM classifier. The results revealed that the overall accuracies of SVM classification of textured images is 88%, while our fusion methodology obtained an accuracy of up to 96%, depending on the size of the data base.

Unsupervised Image Classification through Multisensor Fusion using Fuzzy Class Vector (퍼지 클래스 벡터를 이용하는 다중센서 융합에 의한 무감독 영상분류)

  • 이상훈
    • Korean Journal of Remote Sensing
    • /
    • v.19 no.4
    • /
    • pp.329-339
    • /
    • 2003
  • In this study, an approach of image fusion in decision level has been proposed for unsupervised image classification using the images acquired from multiple sensors with different characteristics. The proposed method applies separately for each sensor the unsupervised image classification scheme based on spatial region growing segmentation, which makes use of hierarchical clustering, and computes iteratively the maximum likelihood estimates of fuzzy class vectors for the segmented regions by EM(expected maximization) algorithm. The fuzzy class vector is considered as an indicator vector whose elements represent the probabilities that the region belongs to the classes existed. Then, it combines the classification results of each sensor using the fuzzy class vectors. This approach does not require such a high precision in spatial coregistration between the images of different sensors as the image fusion scheme of pixel level does. In this study, the proposed method has been applied to multispectral SPOT and AIRSAR data observed over north-eastern area of Jeollabuk-do, and the experimental results show that it provides more correct information for the classification than the scheme using an augmented vector technique, which is the most conventional approach of image fusion in pixel level.

Analysis of Land Cover Changes Based on Classification Result Using PlanetScope Satellite Imagery

  • Yoon, Byunghyun;Choi, Jaewan
    • Korean Journal of Remote Sensing
    • /
    • v.34 no.4
    • /
    • pp.671-680
    • /
    • 2018
  • Compared to the imagery produced by traditional satellites, PlanetScope satellite imagery has made it possible to easily capture remotely-sensed imagery every day through dozens or even hundreds of satellites on a relatively small budget. This study aimed to detect changed areas and update a land cover map using a PlanetScope image. To generate a classification map, pixel-based Random Forest (RF) classification was performed by using additional features, such as the Normalized Difference Water Index (NDWI) and the Normalized Difference Vegetation Index (NDVI). The classification result was converted to vector data and compared with the existing land cover map to estimate the changed area. To estimate the accuracy and trends of the changed area, the quantitative quality of the supervised classification result using the PlanetScope image was evaluated first. In addition, the patterns of the changed area that corresponded to the classification result were analyzed using the PlanetScope satellite image. Experimental results found that the PlanetScope image can be used to effectively to detect changed areas on large-scale land cover maps, and supervised classification results can update the changed areas.

Classification of Objects using CNN-Based Vision and Lidar Fusion in Autonomous Vehicle Environment

  • G.komali ;A.Sri Nagesh
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.11
    • /
    • pp.67-72
    • /
    • 2023
  • In the past decade, Autonomous Vehicle Systems (AVS) have advanced at an exponential rate, particularly due to improvements in artificial intelligence, which have had a significant impact on social as well as road safety and the future of transportation systems. The fusion of light detection and ranging (LiDAR) and camera data in real-time is known to be a crucial process in many applications, such as in autonomous driving, industrial automation and robotics. Especially in the case of autonomous vehicles, the efficient fusion of data from these two types of sensors is important to enabling the depth of objects as well as the classification of objects at short and long distances. This paper presents classification of objects using CNN based vision and Light Detection and Ranging (LIDAR) fusion in autonomous vehicles in the environment. This method is based on convolutional neural network (CNN) and image up sampling theory. By creating a point cloud of LIDAR data up sampling and converting into pixel-level depth information, depth information is connected with Red Green Blue data and fed into a deep CNN. The proposed method can obtain informative feature representation for object classification in autonomous vehicle environment using the integrated vision and LIDAR data. This method is adopted to guarantee both object classification accuracy and minimal loss. Experimental results show the effectiveness and efficiency of presented approach for objects classification.

Study on an algorithm for atmospheric correction of Landsat TM imagery using MODTRAN simulation

  • Oh, Sung-Nam;Yu, Sung-Yeol;Lee, Hyun-Kyung;Kim, Yong-Sup;Park, Kyung-Won
    • Proceedings of the KSRS Conference
    • /
    • 1998.09a
    • /
    • pp.103-109
    • /
    • 1998
  • A technique on atmospheric correction algorithm for a single band (0.76-0.90 $\mu$m) reflective of Landsat TM imagery has been developed using a radiation transfer model simulation. It proceeds in two steps: First, calculation of the surface reflectance of each pixel based on precomputed planetary albedo functions for actual atmospheres(e. g. radiosonde) and two kinds of atmospheric visibility states. Second, approximate correction of the adjacency pixel effect by taking into account the average reflectance in an 7 $\times$ 7 pixel neighbourhood and using appropriate land cover classification in reflectance. The correction functions are provided by MODTRAN model.

  • PDF

Geostatistical Fusion of Spectral and Spatial Information in Remote Sensing Data Classification

  • Park, No-Wook;Chi, Kwang-Hoon;Kwon, Byung-Doo
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.399-401
    • /
    • 2003
  • This paper presents a geostatistical contextual classifier for the classification of remote sensing data. To obtain accurate spatial/contextual information, a simple indicator kriging algorithm with local means that allows one to estimate the probability of occurrence of certain classes on the basis of surrounding pixel information is applied. To illustrate the proposed scheme, supervised classification of multi-sensor remote sensing data is carried out. Analysis of the results indicates that the proposed method improved the classification accuracy, compared to the method based on the spectral information only.

  • PDF

Unsupervised Multispectral Image Segmentation Based on 1D Combined Neighborhood Differences (1D 통합된 근접차이에 기반한 자율적인 다중분광 영상 분할)

  • Saipullah, Khairul Muzzammil;Yun, Byung-Choon;Kim, Deok-Hwan
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2010.11a
    • /
    • pp.625-628
    • /
    • 2010
  • This paper proposes a novel feature extraction method for unsupervised multispectral image segmentation based in one dimensional combined neighborhood differences (1D CND). In contrast with the original CND, which is applied with traditional image, 1D CND is computed on a single pixel with various bands. The proposed algorithm utilizes the sign of differences between bands of the pixel. The difference values are thresholded to form a binary codeword. A binomial factor is assigned to these codeword to form another unique value. These values are then grouped to construct the 1D CND feature image where is used in the unsupervised image segmentation. Various experiments using two LANDSAT multispectral images have been performed to evaluate the segmentation and classification accuracy of the proposed method. The result shows that 1D CND feature outperforms the spectral feature, with average classification accuracy of 87.55% whereas that of spectral feature is 55.81%.

Texture Classification Algorithm for Patch-based Image Processing (패치 기반 영상처리를 위한 텍스쳐 분류 알고리즘)

  • Yu, Seung Wan;Song, Byung Cheol
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.51 no.11
    • /
    • pp.146-154
    • /
    • 2014
  • The local binary pattern (LBP) scheme that is one of the texture classification methods normally uses the distribution of flat, edge and corner patterns. However, it cannot examine the edge direction and the pixel difference because it is a sort of binary pattern caused by thresholding. Furthermore, since it cannot consider the pixel distribution, it shows lower performance as the image size becomes larger. In order to solve this problem, we propose a sub-classification method using the edge direction distribution and eigen-matrix. The proposed sub-classification is applied to the particular texture patches which cannot be classified by LBP. First, we quantize the edge direction and compute its distribution. Second, we calculate the distribution of the largest value among eigenvalues derived from structure matrix. Simulation results show that the proposed method provides a higher classification performance of about 8 % than the existing method.