• Title/Summary/Keyword: classifiers

Search Result 743, Processing Time 0.026 seconds

EEG Dimensional Reduction with Stack AutoEncoder for Emotional Recognition using LSTM/RNN (LSTM/RNN을 사용한 감정인식을 위한 스택 오토 인코더로 EEG 차원 감소)

  • Aliyu, Ibrahim;Lim, Chang-Gyoon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.15 no.4
    • /
    • pp.717-724
    • /
    • 2020
  • Due to the important role played by emotion in human interaction, affective computing is dedicated in trying to understand and regulate emotion through human-aware artificial intelligence. By understanding, emotion mental diseases such as depression, autism, attention deficit hyperactivity disorder, and game addiction will be better managed as they are all associated with emotion. Various studies for emotion recognition have been conducted to solve these problems. In applying machine learning for the emotion recognition, the efforts to reduce the complexity of the algorithm and improve the accuracy are required. In this paper, we investigate emotion Electroencephalogram (EEG) feature reduction and classification using Stack AutoEncoder (SAE) and Long-Short-Term-Memory/Recurrent Neural Networks (LSTM/RNN) classification respectively. The proposed method reduced the complexity of the model and significantly enhance the performance of the classifiers.

A Study on Synthetic Data Generation Based Safe Differentially Private GAN (차분 프라이버시를 만족하는 안전한 GAN 기반 재현 데이터 생성 기술 연구)

  • Kang, Junyoung;Jeong, Sooyong;Hong, Dowon;Seo, Changho
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.30 no.5
    • /
    • pp.945-956
    • /
    • 2020
  • The publication of data is essential in order to receive high quality services from many applications. However, if the original data is published as it is, there is a risk that sensitive information (political tendency, disease, ets.) may reveal. Therefore, many research have been proposed, not the original data but the synthetic data generating and publishing to privacy preserve. but, there is a risk of privacy leakage still even if simply generate and publish the synthetic data by various attacks (linkage attack, inference attack, etc.). In this paper, we propose a synthetic data generation algorithm in which privacy preserved by applying differential privacy the latest privacy protection technique to GAN, which is drawing attention as a synthetic data generative model in order to prevent the leakage of such sensitive information. The generative model used CGAN for efficient learning of labeled data, and applied Rényi differential privacy, which is relaxation of differential privacy, considering the utility aspects of the data. And validation of the utility of the generated data is conducted and compared through various classifiers.

Text Classification based on a Feature Projection Technique with Robustness from Noisy Data (오류 데이타에 강한 자질 투영법 기반의 문서 범주화 기법)

  • 고영중;서정연
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.4
    • /
    • pp.498-504
    • /
    • 2004
  • This paper presents a new text classifier based on a feature projection technique. In feature projections, training documents are represented as the projections on each feature. A classification process is based on individual feature projections. The final classification is determined by the sum from the individual classification of each feature. In our experiments, the proposed classifier showed high performance. Especially, it have fast execution speed and robustness with noisy data in comparison with k-NN and SVM, which are among the state-of-art text classifiers. Since the algorithm of the proposed classifier is very simple, its implementation and training process can be done very simply. Therefore, it can be a useful classifier in text classification tasks which need fast execution speed, robustness, and high performance.

Searching for Optimal Ensemble of Feature-classifier Pairs in Gene Expression Profile using Genetic Algorithm (유전알고리즘을 이용한 유전자발현 데이타상의 특징-분류기쌍 최적 앙상블 탐색)

  • 박찬호;조성배
    • Journal of KIISE:Software and Applications
    • /
    • v.31 no.4
    • /
    • pp.525-536
    • /
    • 2004
  • Gene expression profile is numerical data of gene expression level from organism, measured on the microarray. Generally, each specific tissue indicates different expression levels in related genes, so that we can classify disease with gene expression profile. Because all genes are not related to disease, it is needed to select related genes that is called feature selection, and it is needed to classify selected genes properly. This paper Proposes GA based method for searching optimal ensemble of feature-classifier pairs that are composed with seven feature selection methods based on correlation, similarity, and information theory, and six representative classifiers. In experimental results with leave-one-out cross validation on two gene expression Profiles related to cancers, we can find ensembles that produce much superior to all individual feature-classifier fairs for Lymphoma dataset and Colon dataset.

Recognition of Unconstrained Handwritten Numerals using Modified Chaotic Neural Networks (수정된 카오스 신경망을 이용한 무제약 서체 숫자 인식)

  • 최한고;김상희;이상재
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.2 no.1
    • /
    • pp.44-52
    • /
    • 2001
  • This paper describes an off-line method for recognizing totally unconstrained handwritten digits using modified chaotic neural networks(MCNN). The chaotic neural networks(CNN) is modified to be a useful network for solving complex pattern problems by enforcing dynamic characteristics and learning process. Since the MCNN has the characteristics of highly nonlinear dynamics in structure and neuron itself, it can be an appropriate network for the robust classification of complex handwritten digits. Digit identification starts with extraction of features from the raw digit images and then recognizes digits using the MCNN based classifier. The performance of the MCNN classifier is evaluated on the numeral database of Concordia University, Montreal, Canada. For the relative comparison of recognition performance, the MCNN classifier is compared with the recurrent neural networks(RNN) classifier. Experimental results show that the classification rate is 98.0%. It indicates that the MCNN classifier outperforms the RNN classifier as well as other classifiers that have been reported on the same database.

  • PDF

Smarter Classification for Imbalanced Data Set and Its Application to Patent Evaluation (불균형 데이터 집합에 대한 스마트 분류방법과 특허 평가에의 응용)

  • Kwon, Ohbyung;Lee, Jonathan Sangyun
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.15-34
    • /
    • 2014
  • Overall, accuracy as a performance measure does not fully consider modular accuracy: the accuracy of classifying 1 (or true) as 1 is not same as classifying 0 (or false) as 0. A smarter classification algorithm would optimize the classification rules to match the modular accuracies' goals according to the nature of problem. Correspondingly, smarter algorithms must be both more generalized with respect to the nature of problems, and free from decretization, which may cause distortion of the real performance. Hence, in this paper, we propose a novel vertical boosting algorithm that improves modular accuracies. Rather than decretizing items, we use simple classifiers such as a regression model that accepts continuous data types. To improve the generalization, and to select a classification model that is well-suited to the nature of the problem domain, we developed a model selection algorithm with smartness. To show the soundness of the proposed method, we performed an experiment with a real-world application: predicting the intellectual properties of e-transaction technology, which had a 47,000+ record data set.

Adjustment of Lactation Number and Stage on Informal Linear Type Traits of Holstein Dairy Cattle

  • Do, Chang-Hee;Jeon, Beong-Soon;Sang, Byung-Chan;Lee, Dong-Hee;Pearson, Ronald E.
    • Journal of Animal Science and Technology
    • /
    • v.52 no.6
    • /
    • pp.467-473
    • /
    • 2010
  • A total of 4,323,781 records for informal 16 primary linear descriptive traits of dairy cows in Holstein breed from 1988 to 2007 in USA were analyzed to estimate adjustment factors for lactation number and stage. While all factors in the model were highly significant (P < 0.01), major influences on linear type traits were due to lactation number and stage. The frequencies of lactation number 1 through 6 were 58.6, 22.0, 11.8, 4.8, 2.1, and 0.8%, respectively. Further, the frequencies of lactation stage were 0.7, 76.9, 15.3, 4.9, and 2.1%, respectively, for springing, early, medium, late, and dry. To adjust 16 linear traits (stature, dairy form, strength, body depth, rump width, rump angle, legs rear view, leg set, foot angle, fore udder, rear udder height, rear udder width, udder support, udder depth, and front teat placement), additive and multiplicative adjustment factors of lactation number (lactations 2 to 4) and stage (springing, medium, late and dry) were estimated with the solutions in the generalized linear model, assigning lactation 1 and stage early as base class. Additive adjustment factors of lactation number ranged from -1.23 to 2.908, while multiplicative factors ranged from 0.853 to 2.207. Further, additive and multiplicative adjustment factors for lactation stage ranged from -0.668 to 0.785, and from 0.891 to 1.154. Application of adjustment factors to 20 randomly sampled sub-data sets produced the results that additive adjustment factors for both lactation number and stage reduced more mean square of lactation number and stage over 16 linear traits than any combination of adjustments, and leaded additive adjustment factors for both lactation number and stage as a choice of methods for adjustment of informal 16 primary linear type traits collected by classifiers of AI studs.

An Ensemble Method for Latent Interest Reasoning of Mobile Users (모바일 사용자의 잠재 관심 추론을 위한 앙상블 기법)

  • Choi, Yerim;Park, Jonghun;Shin, Dong Wan
    • KIISE Transactions on Computing Practices
    • /
    • v.21 no.11
    • /
    • pp.706-712
    • /
    • 2015
  • These days, much information is provided as a list of summaries through mobile services. In this regard, users consume information in which they are interested by observing the list and not by expressing their interest explicitly or implicitly through rating content or clicking links. Therefore, to appropriately model a user's interest, it is necessary to detect latent interest content. In this study, we propose a method for reasoning latent interest of a user by analyzing mobile content consumption logs of the user. Specifically, since erroneous reasoning will drastically degrade service quality, a unanimity ensemble method is adopted to maximize precision. In this method, an item is determined as the subject of latent interest only when multiple classifiers considering various aspects of the log unanimously agree. Accurate reasoning of latent interest will contribute to enhancing the quality of personalized services such as interest-based recommendation systems.

A Meta-learning Approach for Building Multi-classifier Systems in a GA-based Inductive Learning Environment (유전 알고리즘 기반 귀납적 학습 환경에서 다중 분류기 시스템의 구축을 위한 메타 학습법)

  • Kim, Yeong-Joon;Hong, Chul-Eui
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.1
    • /
    • pp.35-40
    • /
    • 2015
  • The paper proposes a meta-learning approach for building multi-classifier systems in a GA-based inductive learning environment. In our meta-learning approach, a classifier consists of a general classifier and a meta-classifier. We obtain a meta-classifier from classification results of its general classifier by applying a learning algorithm to them. The role of the meta-classifier is to evaluate the classification result of its general classifier and decide whether to participate into a final decision-making process or not. The classification system draws a decision by combining classification results that are evaluated as correct ones by meta-classifiers. We present empirical results that evaluate the effect of our meta-learning approach on the performance of multi-classifier systems.

Weighted L1-Norm Support Vector Machine for the Classification of Highly Imbalanced Data (불균형 자료의 분류분석을 위한 가중 L1-norm SVM)

  • Kim, Eunkyung;Jhun, Myoungshic;Bang, Sungwan
    • The Korean Journal of Applied Statistics
    • /
    • v.28 no.1
    • /
    • pp.9-21
    • /
    • 2015
  • The support vector machine has been successfully applied to various classification areas due to its flexibility and a high level of classification accuracy. However, when analyzing imbalanced data with uneven class sizes, the classification accuracy of SVM may drop significantly in predicting minority class because the SVM classifiers are undesirably biased toward the majority class. The weighted $L_2$-norm SVM was developed for the analysis of imbalanced data; however, it cannot identify irrelevant input variables due to the characteristics of the ridge penalty. Therefore, we propose the weighted $L_1$-norm SVM, which uses lasso penalty to select important input variables and weights to differentiate the misclassification of data points between classes. We demonstrate the satisfactory performance of the proposed method through simulation studies and a real data analysis.