• Title/Summary/Keyword: kernel-based method

Search Result 474, Processing Time 0.025 seconds

Kernel-based actor-critic approach with applications

  • Chu, Baek-Suk;Jung, Keun-Woo;Park, Joo-Young
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.11 no.4
    • /
    • pp.267-274
    • /
    • 2011
  • Recently, actor-critic methods have drawn significant interests in the area of reinforcement learning, and several algorithms have been studied along the line of the actor-critic strategy. In this paper, we consider a new type of actor-critic algorithms employing the kernel methods, which have recently shown to be very effective tools in the various fields of machine learning, and have performed investigations on combining the actor-critic strategy together with kernel methods. More specifically, this paper studies actor-critic algorithms utilizing the kernel-based least-squares estimation and policy gradient, and in its critic's part, the study uses a sliding-window-based kernel least-squares method, which leads to a fast and efficient value-function-estimation in a nonparametric setting. The applicability of the considered algorithms is illustrated via a robot locomotion problem and a tunnel ventilation control problem.

Kernel Analysis of Weighted Linear Interpolation Based on Even-Odd Decomposition (짝수 홀수 분해 기반의 가중 선형 보간법을 위한 커널 분석)

  • Oh, Eun-ju;Yoo, Hoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.11
    • /
    • pp.1455-1461
    • /
    • 2018
  • This paper presents a kernel analysis of weighted linear interpolation based on even-odd decomposition (EOD). The EOD method has advantages in that it provides low-complexity and improved image quality than the CCI method. However, since the kernel of EOD has not studied before and its analysis has not been addressed yet, this paper proposes the kernel function and its analysis. The kernel function is divided into odd and even terms. And then, the kernel is accomplished by summing the two terms. The proposed kernel is adjustable by a parameter. The parameter influences efficiency in the EOD based WLI process. Also, the kernel shapes are proposed by adjusting the parameter. In addition, the discussion with respect to the parameter is given to understand the parameter. A preliminary experiment on the kernel shape is presented to understand the adjustable parameter and corresponding kernel.

AN ELIGIBLE PRIMAL-DUAL INTERIOR-POINT METHOD FOR LINEAR OPTIMIZATION

  • Cho, Gyeong-Mi;Lee, Yong-Hoon
    • East Asian mathematical journal
    • /
    • v.29 no.3
    • /
    • pp.279-292
    • /
    • 2013
  • It is well known that each kernel function defines a primal-dual interior-point method(IPM). Most of polynomial-time interior-point algorithms for linear optimization(LO) are based on the logarithmic kernel function([2, 11]). In this paper we define a new eligible kernel function and propose a new search direction and proximity function based on this function for LO problems. We show that the new algorithm has ${\mathcal{O}}((log\;p){\sqrt{n}}\;log\;n\;log\;{\frac{n}{\epsilon}})$ and ${\mathcal{O}}((q\;log\;p)^{\frac{3}{2}}{\sqrt{n}}\;log\;{\frac{n}{\epsilon}})$ iteration bound for large- and small-update methods, respectively. These are currently the best known complexity results.

Centroid Neural Network with Bhattacharyya Kernel (Bhattacharyya 커널을 적용한 Centroid Neural Network)

  • Lee, Song-Jae;Park, Dong-Chul
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.32 no.9C
    • /
    • pp.861-866
    • /
    • 2007
  • A clustering algorithm for Gaussian Probability Distribution Function (GPDF) data called Centroid Neural Network with a Bhattacharyya Kernel (BK-CNN) is proposed in this paper. The proposed BK-CNN is based on the unsupervised competitive Centroid Neural Network (CNN) and employs a kernel method for data projection. The kernel method adopted in the proposed BK-CNN is used to project data from the low dimensional input feature space into higher dimensional feature space so as the nonlinear problems associated with input space can be solved linearly in the feature space. In order to cluster the GPDF data, the Bhattacharyya kernel is used to measure the distance between two probability distributions for data projection. With the incorporation of the kernel method, the proposed BK-CNN is capable of dealing with nonlinear separation boundaries and can successfully allocate more code vector in the region that GPDF data are densely distributed. When applied to GPDF data in an image classification probleml, the experiment results show that the proposed BK-CNN algorithm gives 1.7%-4.3% improvements in average classification accuracy over other conventional algorithm such as k-means, Self-Organizing Map (SOM) and CNN algorithms with a Bhattacharyya distance, classed as Bk-Means, B-SOM, B-CNN algorithms.

Design of an Optimized GPGPU for Data Reuse in DeepLearning Convolution (딥러닝 합성곱에서 데이터 재사용에 최적화된 GPGPU 설계)

  • Nam, Ki-Hun;Lee, Kwang-Yeob;Jung, Jun-Mo
    • Journal of IKEEE
    • /
    • v.25 no.4
    • /
    • pp.664-671
    • /
    • 2021
  • This paper proposes a GPGPU structure that can reduce the number of operations and memory access by effectively applying a data reuse method to a convolutional neural network(CNN). Convolution is a two-dimensional operation using kernel and input data, and the operation is performed by sliding the kernel. In this case, a reuse method using an internal register is proposed instead of loading kernel from a cache memory until the convolution operation is completed. The serial operation method was applied to the convolution to increase the effect of data reuse by using the principle of GPGPU in which instructions are executed by the SIMT method. In this paper, for register-based data reuse, the kernel was fixed at 4×4 and GPGPU was designed considering the warp size and register bank to effectively support it. To verify the performance of the designed GPGPU on the CNN, we implemented it as an FPGA and then ran LeNet and measured the performance on AlexNet by comparison using TensorFlow. As a result of the measurement, 1-iteration learning speed based on AlexNet is 0.468sec and the inference speed is 0.135sec.

Improvement in Supervector Linear Kernel SVM for Speaker Identification Using Feature Enhancement and Training Length Adjustment (특징 강화 기법과 학습 데이터 길이 조절에 의한 Supervector Linear Kernel SVM 화자식별 개선)

  • So, Byung-Min;Kim, Kyung-Wha;Kim, Min-Seok;Yang, Il-Ho;Kim, Myung-Jae;Yu, Ha-Jin
    • The Journal of the Acoustical Society of Korea
    • /
    • v.30 no.6
    • /
    • pp.330-336
    • /
    • 2011
  • In this paper, we propose a new method to improve the performance of supervector linear kernel SVM (Support Vector Machine) for speaker identification. This method is based on splitting one training datum into several pieces of utterances. We use four different databases for evaluating performance and use PCA (Principal Component Analysis), GKPCA (Greedy Kernel PCA) and KMDA (Kernel Multimodal Discriminant Analysis) for feature enhancement. As a result, the proposed method shows improved performance for speaker identification using supervector linear kernel SVM.

GMM Based Voice Conversion Using Kernel PCA (Kernel PCA를 이용한 GMM 기반의 음성변환)

  • Han, Joon-Hee;Bae, Jae-Hyun;Oh, Yung-Hwan
    • MALSORI
    • /
    • no.67
    • /
    • pp.167-180
    • /
    • 2008
  • This paper describes a novel spectral envelope conversion method based on Gaussian mixture model (GMM). The core of this paper is rearranging source feature vectors in input space to the transformed feature vectors in feature space for the better modeling of GMM of source and target features. The quality of statistical modeling is dependent on the distribution and the dimension of data. The proposed method transforms both of the distribution and dimension of data and gives us the chance to model the same data with different configuration. Because the converted feature vectors should be on the input space, only source feature vectors are rearranged in the feature space and target feature vectors remain unchanged for the joint pdf of source and target features using KPCA. The experimental result shows that the proposed method outperforms the conventional GMM-based conversion method in various training environment.

  • PDF

A fast gamma-ray dose rate assessment method for complex geometries based on stylized model reconstruction

  • Yang, Li-qun;Liu, Yong-kuo;Peng, Min-jun;Li, Meng-kun;Chao, Nan
    • Nuclear Engineering and Technology
    • /
    • v.51 no.5
    • /
    • pp.1436-1443
    • /
    • 2019
  • A fast gamma-ray dose rate assessment method for complex geometries based on stylized model reconstruction and point-kernel method is proposed in this paper. The complex three-dimensional (3D) geometries are imported as a 3DS format file from 3dsMax software with material and radiometric attributes. Based on 3D stylized model reconstruction of solid mesh, the 3D-geometrical solids are automatically converted into stylized models. In point-kernel calculation, the stylized source models are divided into point kernels and the mean free paths (mfp) are calculated by the intersections between shield stylized models and tracing ray. Compared with MCNP, the proposed method can implement complex 3D geometries visually, and the dose rate calculation is accurate and fast.

Weighted Kernel and it's Learning Method for Cancer Diagnosis System (암진단시스템을 위한 Weighted Kernel 및 학습방법)

  • Choi, Gyoo-Seok;Park, Jong-Jin;Jeon, Byoung-Chan;Park, In-Kyu;Ahn, Ihn-Seok;Nguyen, Ha-Nam
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.9 no.2
    • /
    • pp.1-6
    • /
    • 2009
  • One of the most important problems in bioinformatics is how to extract the useful information from a huge amount of data, and make a decision in diagnosis, prognosis, and medical treatment applications. This paper proposes a weighted kernel function for support vector machine and its learning method with a fast convergence and a good classification performance. We defined the weighted kernel function as the weighted sum of a set of different types of basis kernel functions such as neural, radial, and polynomial kernels, which are trained by a learning method based on genetic algorithm. The weights of basis kernel functions in proposed kernel are determined in learning phase and used as the parameters in the decision model in classification phase. The experiments on several clinical datasets such as colon cancer indicate that our weighted kernel function results in higher and more stable classification performance than other kernel functions.

  • PDF

Variable Bandwidth Selection for Kernel Regression

  • Kim, Dae-Hak
    • Journal of the Korean Data and Information Science Society
    • /
    • v.5 no.1
    • /
    • pp.11-20
    • /
    • 1994
  • In recent years, nonparametric kernel estimation of regresion function are abundant and widely applicable to many areas of statistics. Most of modern researches concerned with the fixed global bandwidth selection which can be used in the estimation of regression function with all the same value for all x. In this paper, we propose a method for selecting locally varing bandwidth based on bootstrap method in kernel estimation of fixed design regression. Performance of proposed bandwidth selection method for finite sample case is conducted via Monte Carlo simulation study.

  • PDF