• Title/Summary/Keyword: Nonlinear principal component analysis

Search Result 57, Processing Time 0.02 seconds

Arrow Diagrams for Kernel Principal Component Analysis

  • Huh, Myung-Hoe
    • Communications for Statistical Applications and Methods
    • /
    • v.20 no.3
    • /
    • pp.175-184
    • /
    • 2013
  • Kernel principal component analysis(PCA) maps observations in nonlinear feature space to a reduced dimensional plane of principal components. We do not need to specify the feature space explicitly because the procedure uses the kernel trick. In this paper, we propose a graphical scheme to represent variables in the kernel principal component analysis. In addition, we propose an index for individual variables to measure the importance in the principal component plane.

Blind Source Separation via Principal Component Analysis

  • Choi, Seung-Jin
    • Journal of KIEE
    • /
    • v.11 no.1
    • /
    • pp.1-7
    • /
    • 2001
  • Various methods for blind source separation (BSS) are based on independent component analysis (ICA) which can be viewed as a nonlinear extension of principal component analysis (PCA). Most existing ICA methods require certain nonlinear functions (which leads to higher-order statistics) depending on the probability distributions of sources, whereas PCA is a linear learning method based on second-order statistics. In this paper we show that the PCA can be applied to the task of BBS, provided that source are spatially uncorrelated but temporally correlated. Since the resulting method is based on only second-order statistics, it avoids the nonlinear function and is able to separate mixtures of several colored Gaussian sources, in contrast to the conventional ICA methods.

  • PDF

Nonlinear Feature Extraction using Class-augmented Kernel PCA (클래스가 부가된 커널 주성분분석을 이용한 비선형 특징추출)

  • Park, Myoung-Soo;Oh, Sang-Rok
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.48 no.5
    • /
    • pp.7-12
    • /
    • 2011
  • In this papwer, we propose a new feature extraction method, named as Class-augmented Kernel Principal Component Analysis (CA-KPCA), which can extract nonlinear features for classification. Among the subspace method that was being widely used for feature extraction, Class-augmented Principal Component Analysis (CA-PCA) is a recently one that can extract features for a accurate classification without computational difficulties of other methods such as Linear Discriminant Analysis (LDA). However, the features extracted by CA-PCA is still restricted to be in a linear subspace of the original data space, which limites the use of this method for various problems requiring nonlinear features. To resolve this limitation, we apply a kernel trick to develop a new version of CA-PCA to extract nonlinear features, and evaluate its performance by experiments using data sets in the UCI Machine Learning Repository.

Incremental Eigenspace Model Applied To Kernel Principal Component Analysis

  • Kim, Byung-Joo
    • Journal of the Korean Data and Information Science Society
    • /
    • v.14 no.2
    • /
    • pp.345-354
    • /
    • 2003
  • An incremental kernel principal component analysis(IKPCA) is proposed for the nonlinear feature extraction from the data. The problem of batch kernel principal component analysis(KPCA) is that the computation becomes prohibitive when the data set is large. Another problem is that, in order to update the eigenvectors with another data, the whole eigenvectors should be recomputed. IKPCA overcomes this problem by incrementally updating the eigenspace model. IKPCA is more efficient in memory requirement than a batch KPCA and can be easily improved by re-learning the data. In our experiments we show that IKPCA is comparable in performance to a batch KPCA for the classification problem on nonlinear data set.

  • PDF

Abnormality Detection to Non-linear Multivariate Process Using Supervised Learning Methods (지도학습기법을 이용한 비선형 다변량 공정의 비정상 상태 탐지)

  • Son, Young-Tae;Yun, Deok-Kyun
    • IE interfaces
    • /
    • v.24 no.1
    • /
    • pp.8-14
    • /
    • 2011
  • Principal Component Analysis (PCA) reduces the dimensionality of the process by creating a new set of variables, Principal components (PCs), which attempt to reflect the true underlying process dimension. However, for highly nonlinear processes, this form of monitoring may not be efficient since the process dimensionality can't be represented by a small number of PCs. Examples include the process of semiconductors, pharmaceuticals and chemicals. Nonlinear correlated process variables can be reduced to a set of nonlinear principal components, through the application of Kernel Principal Component Analysis (KPCA). Support Vector Data Description (SVDD) which has roots in a supervised learning theory is a training algorithm based on structural risk minimization. Its control limit does not depend on the distribution, but adapts to the real data. So, in this paper proposes a non-linear process monitoring technique based on supervised learning methods and KPCA. Through simulated examples, it has been shown that the proposed monitoring chart is more effective than $T^2$ chart for nonlinear processes.

On-line Nonlinear Principal Component Analysis for Nonlinear Feature Extraction (비선형 특징 추출을 위한 온라인 비선형 주성분분석 기법)

  • 김병주;심주용;황창하;김일곤
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.3
    • /
    • pp.361-368
    • /
    • 2004
  • The purpose of this study is to propose a new on-line nonlinear PCA(OL-NPCA) method for a nonlinear feature extraction from the incremental data. Kernel PCA(KPCA) is widely used for nonlinear feature extraction, however, it has been pointed out that KPCA has the following problems. First, applying KPCA to N patterns requires storing and finding the eigenvectors of a N${\times}$N kernel matrix, which is infeasible for a large number of data N. Second problem is that in order to update the eigenvectors with an another data, the whole eigenspace should be recomputed. OL-NPCA overcomes these problems by incremental eigenspace update method with a feature mapping function. According to the experimental results, which comes from applying OL-NPCA to a toy and a large data problem, OL-NPCA shows following advantages. First, OL-NPCA is more efficient in memory requirement than KPCA. Second advantage is that OL-NPCA is comparable in performance to KPCA. Furthermore, performance of OL-NPCA can be easily improved by re-learning the data.

A Comparative Study on Isomap-based Damage Localization (아이소맵을 이용한 결함 탐지 비교 연구)

  • Koh, Bong-Hwan;Jeong, Min-Joong
    • Proceedings of the Computational Structural Engineering Institute Conference
    • /
    • 2011.04a
    • /
    • pp.278-281
    • /
    • 2011
  • The global coordinates generated from Isomap algorithm provide a simple way to analyze and manipulate high dimensional observations in terms of their intrinsic nonlinear degrees of freedom. Thus, Isomap can find globally meaningful coordinates and nonlinear structure of complex data sets, while neither principal component analysis (PCA) nor multidimensional scaling (MDS) are successful in many cases. It is demonstrated that the adapted Isomap algorithm successfully enhances the quality of pattern classification for damage identification in various numerical examples.

  • PDF

An efficient learning algorithm of nonlinear PCA neural networks using momentum (모멘트를 이용한 비선형 주요성분분석 신경망의 효율적인 학습알고리즘)

  • Cho, Yong-Hyun
    • Journal of the Korean Society of Industry Convergence
    • /
    • v.3 no.4
    • /
    • pp.361-367
    • /
    • 2000
  • This paper proposes an efficient feature extraction of the image data using nonlinear principal component analysis neural networks of a new learning algorithm. The proposed method is a learning algorithm with momentum for reflecting the past trends. It is to get the better performance by restraining an oscillation due to converge the global optimum. The proposed algorithm has been applied to the cancer image of $256{\times}256$ pixels and the coin image of $128{\times}128$ pixels respectively. The simulation results show that the proposed algorithm has better performances of the convergence and the nonlinear feature extraction, in comparison with those using the backpropagation and the conventional nonlinear PCA neural networks.

  • PDF

The Kernel Trick for Content-Based Media Retrieval in Online Social Networks

  • Cha, Guang-Ho
    • Journal of Information Processing Systems
    • /
    • v.17 no.5
    • /
    • pp.1020-1033
    • /
    • 2021
  • Nowadays, online or mobile social network services (SNS) are very popular and widely spread in our society and daily lives to instantly share, disseminate, and search information. In particular, SNS such as YouTube, Flickr, Facebook, and Amazon allow users to upload billions of images or videos and also provide a number of multimedia information to users. Information retrieval in multimedia-rich SNS is very useful but challenging task. Content-based media retrieval (CBMR) is the process of obtaining the relevant image or video objects for a given query from a collection of information sources. However, CBMR suffers from the dimensionality curse due to inherent high dimensionality features of media data. This paper investigates the effectiveness of the kernel trick in CBMR, specifically, the kernel principal component analysis (KPCA) for dimensionality reduction. KPCA is a nonlinear extension of linear principal component analysis (LPCA) to discovering nonlinear embeddings using the kernel trick. The fundamental idea of KPCA is mapping the input data into a highdimensional feature space through a nonlinear kernel function and then computing the principal components on that mapped space. This paper investigates the potential of KPCA in CBMR for feature extraction or dimensionality reduction. Using the Gaussian kernel in our experiments, we compute the principal components of an image dataset in the transformed space and then we use them as new feature dimensions for the image dataset. Moreover, KPCA can be applied to other many domains including CBMR, where LPCA has been used to extract features and where the nonlinear extension would be effective. Our results from extensive experiments demonstrate that the potential of KPCA is very encouraging compared with LPCA in CBMR.

Application of Sensor Fault Detection Scheme Based on AANN to Sensor Network (AANN-기반 센서 고장 검출 기법의 센서 네트워크에의 적용)

  • Lee, Young-Sam;Kim, Sung-Ho
    • Proceedings of the KIEE Conference
    • /
    • 2006.10c
    • /
    • pp.229-231
    • /
    • 2006
  • NLPCA(Nonlinear Principal Component Analysis) is a novel technique for multivariate data analysis, similar to the well-known method of principal component analysis. NLPCA operates by a feedforward neural network called AANN(Auto Associative Neural Network) which performs the identity mapping. In this work, a sensor fault detection system based on NLPCA is presented. To verify its applicability, simulation study on the data supplied from sensor network is executed.

  • PDF