• Title/Summary/Keyword: Nearest

Search Result 1,771, Processing Time 0.027 seconds

On the Use of Modified Adaptive Nearest Neighbors for Classification (수정된 적응 최근접 방법을 활용한 판별분류방법에 대한 연구)

  • Maeng, Jin-Woo;Bang, Sung-Wan;Jhun, Myoung-Shic
    • The Korean Journal of Applied Statistics
    • /
    • v.23 no.6
    • /
    • pp.1093-1102
    • /
    • 2010
  • Even though the k-Nearest Neighbors Classification(KNNC) is one of the popular non-parametric classification methods, it does not consider the local features and class information for each observation. In order to overcome such limitations, several methods have been developed such as Adaptive Nearest Neighbors Classification(ANNC) and Modified k-Nearest Neighbors Classification(MKNNC). In this paper, we propose the Modified Adaptive Nearest Neighbors Classification(MANNC) that employs the advantages of both the ANNC and MKNNC. Through a real data analysis and a simulation study, we show that the proposed MANNC outperforms other methods in terms of classification accuracy.

On the use of weighted adaptive nearest neighbors for missing value imputation (가중 적응 최근접 이웃을 이용한 결측치 대치)

  • Yum, Yunjin;Kim, Dongjae
    • The Korean Journal of Applied Statistics
    • /
    • v.31 no.4
    • /
    • pp.507-516
    • /
    • 2018
  • Widely used among the various single imputation methods is k-nearest neighbors (KNN) imputation due to its robustness even when a parametric model such as multivariate normality is not satisfied. We propose a weighted adaptive nearest neighbors imputation method that combines the adaptive nearest neighbors imputation method that accounts for the local features of the data in the KNN imputation method and weighted k-nearest neighbors method that are less sensitive to extreme value or outlier among k-nearest neighbors. We conducted a Monte Carlo simulation study to compare the performance of the proposed imputation method with previous imputation methods.

On the Use of Sequential Adaptive Nearest Neighbors for Missing Value Imputation (순차 적응 최근접 이웃을 활용한 결측값 대치법)

  • Park, So-Hyun;Bang, Sung-Wan;Jhun, Myoung-Shic
    • The Korean Journal of Applied Statistics
    • /
    • v.24 no.6
    • /
    • pp.1249-1257
    • /
    • 2011
  • In this paper, we propose a Sequential Adaptive Nearest Neighbor(SANN) imputation method that combines the Adaptive Nearest Neighbor(ANN) method and the Sequential k-Nearest Neighbor(SKNN) method. When choosing the nearest neighbors of missing observations, the proposed SANN method takes the local feature of the missing observations into account as well as reutilizes the imputed observations in a sequential manner. By using a Monte Carlo study and a real data example, we demonstrate the characteristics of the SANN method and its potential performance.

The privacy protection algorithm of ciphertext nearest neighbor query based on the single Hilbert curve

  • Tan, Delin;Wang, Huajun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.9
    • /
    • pp.3087-3103
    • /
    • 2022
  • Nearest neighbor query in location-based services has become a popular application. Aiming at the shortcomings of the privacy protection algorithms of traditional ciphertext nearest neighbor query having the high system overhead because of the usage of the double Hilbert curves and having the inaccurate query results in some special circumstances, a privacy protection algorithm of ciphertext nearest neighbor query which is based on the single Hilbert curve has been proposed. This algorithm uses a single Hilbert curve to transform the two-dimensional coordinates of the points of interest into Hilbert values, and then encrypts them by the order preserving encryption scheme to obtain the one-dimensional ciphertext data which can be compared in numerical size. Then stores the points of interest as elements composed of index value and the ciphertext of the other information about the points of interest on the server-side database. When the user needs to use the nearest neighbor query, firstly calls the approximate nearest neighbor query algorithm proposed in this paper to query on the server-side database, and then obtains the approximate nearest neighbor query results. After that, the accurate nearest neighbor query result can be obtained by calling the precision processing algorithm proposed in this paper. The experimental results show that this privacy protection algorithm of ciphertext nearest neighbor query which is based on the single Hilbert curve is not only feasible, but also optimizes the system overhead and the accuracy of ciphertext nearest neighbor query result.

Adaptive Nearest Neighbors를 활용한 결측치 대치

  • 전명식;정형철
    • Proceedings of the Korean Statistical Society Conference
    • /
    • 2004.11a
    • /
    • pp.185-190
    • /
    • 2004
  • 비모수적 결측치 대치 방법으로 널리 사용되는 k-nearest neighbors(KNN) 방법은 자료의 국소적(local) 특징을 고려하지 않고 전체 자료에 대해 균일한 이웃의 개수 k를 사용하는 단점이 있다. 본 연구에서는 KNN의 대안으로 자료의 국소적 특징을 고려하는 adaptive nearest neighbors(ANN) 방법을 제안하였다. 나아가 microarray 자료의 경우에 대하여 결측치 대치를 통해 KNN과 ANN의 성능을 비교하였다.

  • PDF

Design of Nearest Prototype Classifier by using Differential Evolutionary Algorithm (차분진화 알고리즘을 이용한 Nearest Prototype Classifier 설계)

  • Roh, Seok-Beom;Ahn, Tae-Chon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.4
    • /
    • pp.487-492
    • /
    • 2011
  • In this paper, we proposed a new design methodology to improve the classification performance of the Nearest Prototype Classifier which is one of the simplest classification algorithm. To optimize the position vectors of the prototypes in the nearest prototype classifier, we use the differential evolutionary algorithm. The optimized position vectors of the prototypes result in the improvement of the classification performance. The new method to determine the class labels of the prototypes, which are defined by the differential evolutionary algorithm, is proposed. In addition, the experimental application covers a comparative analysis including several previously commonly encountered methods.

Locality-Sensitive Hashing Techniques for Nearest Neighbor Search

  • Lee, Keon Myung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.12 no.4
    • /
    • pp.300-307
    • /
    • 2012
  • When the volume of data grows big, some simple tasks could become a significant concern. Nearest neighbor search is such a task which finds from a data set the k nearest data points to queries. Locality-sensitive hashing techniques have been developed for approximate but fast nearest neighbor search. This paper introduces the notion of locality-sensitive hashing and surveys the locality-sensitive hashing techniques. It categories them based on several criteria, presents their characteristics, and compares their performance.

A probabilistic nearest neighbor filter incorporating numbers of validated measurements

  • Sang J. Shin;Song, Taek-Lyul;Ahn, Jo-Young
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2002.10a
    • /
    • pp.82.1-82
    • /
    • 2002
  • $\textbullet$ Nearest neighbor filter $\textbullet$ Probabilistic nearest neighbor filter $\textbullet$ Probabilistic nearest neighbor filter incorporating numbers of validated measurements $\textbullet$ Probability density function of the NDS $\textbullet$ Simulation results in a clutter environment to verify the performances $\textbullet$ Sensitivity analysis for the unknown spatial clutter density

  • PDF

Adaptive Nearest Neighbors for Classification (Adaptive Nearest Neighbors를 활용한 판별분류방법)

  • Jhun, Myoung-Shic;Choi, In-Kyung
    • The Korean Journal of Applied Statistics
    • /
    • v.22 no.3
    • /
    • pp.479-488
    • /
    • 2009
  • The ${\kappa}$-Nearest Neighbors Classification(KNNC) is a popular non-parametric classification method which assigns a fixed number ${\kappa}$ of neighbors to every observation without consideration of the local feature of the each observation. In this paper, we propose an Adaptive Nearest Neighbors Classification(ANNC) as an alternative to KNNC. The proposed ANNC method adapts the number of neighbors according to the local feature of the observation such as density of data. To verify characteristics of ANNC, we compare the number of misclassified observation with KNNC by Monte Carlo study and confirm the potential performance of ANNC method.

On the Use of Weighted k-Nearest Neighbors for Missing Value Imputation (Weighted k-Nearest Neighbors를 이용한 결측치 대치)

  • Lim, Chanhui;Kim, Dongjae
    • The Korean Journal of Applied Statistics
    • /
    • v.28 no.1
    • /
    • pp.23-31
    • /
    • 2015
  • A conventional missing value problem in the statistical analysis k-Nearest Neighbor(KNN) method are used for a simple imputation method. When one of the k-nearest neighbors is an extreme value or outlier, the KNN method can create a bias. In this paper, we propose a Weighted k-Nearest Neighbors(WKNN) imputation method that can supplement KNN's faults. A Monte-Carlo simulation study is also adapted to compare the WKNN method and KNN method using real data set.