• Title/Summary/Keyword: Bayes Net

Search Result 14, Processing Time 0.022 seconds

A Study of Data Mining Methodology for Effective Analysis of False Alarm Event on Mechanical Security System (기계경비시스템 오경보 이벤트 분석을 위한 데이터마이닝 기법 연구)

  • Kim, Jong-Min;Choi, Kyong-Ho;Lee, Dong-Hwi
    • Convergence Security Journal
    • /
    • v.12 no.2
    • /
    • pp.61-70
    • /
    • 2012
  • The objective of this study is to achieve the most optimal data mining for effective analysis of false alarm event on mechanical security system. To perform this, this study searches the cause of false alarm and suggests the data conversion and analysis methods to apply to several algorithm of WEKA, which is a data mining program, based on statistical data for the number of case on movement by false alarm, false alarm rate and cause of false alarm. Analysis methods are used to estimate false alarm and set more effective reaction for false alarm by applying several algorithm. To use the suitable data for effective analysis of false alarm event on mechanical security analysis this study uses Decision Tree, Naive Bayes, BayesNet Apriori and J48Tree algorithm, and applies the algorithm by deducting the highest value.

Crack segmentation in high-resolution images using cascaded deep convolutional neural networks and Bayesian data fusion

  • Tang, Wen;Wu, Rih-Teng;Jahanshahi, Mohammad R.
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.221-235
    • /
    • 2022
  • Manual inspection of steel box girders on long span bridges is time-consuming and labor-intensive. The quality of inspection relies on the subjective judgements of the inspectors. This study proposes an automated approach to detect and segment cracks in high-resolution images. An end-to-end cascaded framework is proposed to first detect the existence of cracks using a deep convolutional neural network (CNN) and then segment the crack using a modified U-Net encoder-decoder architecture. A Naïve Bayes data fusion scheme is proposed to reduce the false positives and false negatives effectively. To generate the binary crack mask, first, the original images are divided into 448 × 448 overlapping image patches where these image patches are classified as cracks versus non-cracks using a deep CNN. Next, a modified U-Net is trained from scratch using only the crack patches for segmentation. A customized loss function that consists of binary cross entropy loss and the Dice loss is introduced to enhance the segmentation performance. Additionally, a Naïve Bayes fusion strategy is employed to integrate the crack score maps from different overlapping crack patches and to decide whether a pixel is crack or not. Comprehensive experiments have demonstrated that the proposed approach achieves an 81.71% mean intersection over union (mIoU) score across 5 different training/test splits, which is 7.29% higher than the baseline reference implemented with the original U-Net.

Traffic Classification Using Machine Learning Algorithms in Practical Network Monitoring Environments (실제 네트워크 모니터링 환경에서의 ML 알고리즘을 이용한 트래픽 분류)

  • Jung, Kwang-Bon;Choi, Mi-Jung;Kim, Myung-Sup;Won, Young-J.;Hong, James W.
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.33 no.8B
    • /
    • pp.707-718
    • /
    • 2008
  • The methodology of classifying traffics is changing from payload based or port based to machine learning based in order to overcome the dynamic changes of application's characteristics. However, current state of traffic classification using machine learning (ML) algorithms is ongoing under the offline environment. Specifically, most of the current works provide results of traffic classification using cross validation as a test method. Also, they show classification results based on traffic flows. However, these traffic classification results are not useful for practical environments of the network traffic monitoring. This paper compares the classification results using cross validation with those of using split validation as the test method. Also, this paper compares the classification results based on flow to those based on bytes. We classify network traffics by using various feature sets and machine learning algorithms such as J48, REPTree, RBFNetwork, Multilayer perceptron, BayesNet, and NaiveBayes. In this paper, we find the best feature sets and the best ML algorithm for classifying traffics using the split validation.

Performance Analysis of Mulitilayer Neural Net Claddifiers Using Simulated Pattern-Generating Processes (모의 패턴생성 프로세스를 이용한 다단신경망분류기의 성능분석)

  • Park, Dong-Seon
    • The Transactions of the Korea Information Processing Society
    • /
    • v.4 no.2
    • /
    • pp.456-464
    • /
    • 1997
  • We describe a random prcess model that prvides sets of patterms whth prcisely contrlolled within-class varia-bility and between-class distinctions.We used these pattems in a simulation study wity the back-propagation netwoek to chracterize its perfotmance as we varied the process-controlling parameters,the statistical differences between the processes,and the random noise on the patterns.Our results indicated that grneralized statistical difference between the processes genrating the patterns provided a good predictor of the difficulty of the clssi-fication problem. Also we analyzed the performance of the Bayes classifier whith the maximum-likeihood cri-terion and we compared the performance of the neural network to that of the Bayes classifier.We found that the performance of neural network was intermediate between that of the simulated and theoretical Bayes classifier.

  • PDF

An Automatic Document Classification with Bayesian Learning (베이지안 학습을 이용한 문서의 자동분류)

  • Kim, Jin-Sang;Shin, Yang-Kyu
    • Journal of the Korean Data and Information Science Society
    • /
    • v.11 no.1
    • /
    • pp.19-30
    • /
    • 2000
  • As the number of online documents increases enormously with the expansion of information technology, the importance of automatic document classification is greatly enlarged. In this paper, an automatic document classification method is investigated and applied to UseNet 20 newsgroup articles to test its efficacy. The classification system uses Naive Bayes classification algorithm and the experimental result shows that a randomly selected newsgroup arcicle can be classified into its own category over 77% accuracy.

  • PDF

Performance Analysis of Machine Learning Algorithms for Application Traffic Classification (애플리케이션 트래픽 분류를 위한 머신러닝 알고리즘 성능 분석)

  • Kim, Sung-Yun;Kim, Myung-Sup
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2008.05a
    • /
    • pp.968-970
    • /
    • 2008
  • 기존에 트래픽 분류 방법으로 payload 분석이나 well-known port를 이용한 방법을 많이 사용했다. 하지만 동적으로 변하는 애플리케이션이 늘어남에 따라 기존 방법으로 애플리케이션 트래픽 분류가 어렵다. 이러한 문제의 대안으로 Machine Learning(ML) 알고리즘을 이용한 애플리케이션 트래픽 분류방법이 연구되고 있다. 기존의 논문에서는 일정 시간동안 수집한 data set을 사용하기 때문에 적게 발생한 애플리케이션은 제대로 분류하지 못하여도 전체적으로는 좋은 성능을 보일 수 있다. 본 논문에서는 이러한 문제를 해결하기 위해 각 애플리케이션마다 동일한 수의 data set을 수집하여 애플리케이션 트래픽을 분류하는 방법을 제시한다. ML 알고리즘 중 J48, REPTree, BayesNet, NaiveBayes, Multilayer Perceptron 알고리즘을 이용하여 애플리케이션 트래픽 분류의 정확도를 비교한다.

IMPLEMENTATION OF DATA ASSIMILATION METHODOLOGY FOR PHYSICAL MODEL UNCERTAINTY EVALUATION USING POST-CHF EXPERIMENTAL DATA

  • Heo, Jaeseok;Lee, Seung-Wook;Kim, Kyung Doo
    • Nuclear Engineering and Technology
    • /
    • v.46 no.5
    • /
    • pp.619-632
    • /
    • 2014
  • The Best Estimate Plus Uncertainty (BEPU) method has been widely used to evaluate the uncertainty of a best-estimate thermal hydraulic system code against a figure of merit. This uncertainty is typically evaluated based on the physical model's uncertainties determined by expert judgment. This paper introduces the application of data assimilation methodology to determine the uncertainty bands of the physical models, e.g., the mean value and standard deviation of the parameters, based upon the statistical approach rather than expert judgment. Data assimilation suggests a mathematical methodology for the best estimate bias and the uncertainties of the physical models which optimize the system response following the calibration of model parameters and responses. The mathematical approaches include deterministic and probabilistic methods of data assimilation to solve both linear and nonlinear problems with the a posteriori distribution of parameters derived based on Bayes' theorem. The inverse problem was solved analytically to obtain the mean value and standard deviation of the parameters assuming Gaussian distributions for the parameters and responses, and a sampling method was utilized to illustrate the non-Gaussian a posteriori distributions of parameters. SPACE is used to demonstrate the data assimilation method by determining the bias and the uncertainty bands of the physical models employing Bennett's heated tube test data and Becker's post critical heat flux experimental data. Based on the results of the data assimilation process, the major sources of the modeling uncertainties were identified for further model development.

A comparison of activity recognition using a triaxial accelerometer sensor (3축 가속도 센서를 이용한 행동 인식 비교)

  • Wang, ChangWon;Ho, JongGab;Na, YeJi;Jung, HwaYung;Nam, YunYoung;Min, Se Dong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2015.10a
    • /
    • pp.1361-1364
    • /
    • 2015
  • 본 연구에서는 노인들이 일상에서 많이 행동하는 7가지 유형의 행동의 특징을 추출하고, 총 7가지 분류 알고리즘에 적용하여 가장 인식률이 높은 알고리즘을 도출하고자 하였다. 행동패턴은 정상보행, 절름발이, 지팡이, 느린 보행, 허리가 굽은 상태에서 보행, 스스로 휠체어 끌 때 그리고 누군가가 휠체어를 끌어줄 때 총 7가지로 구성하였다. 행동패턴의 특징은 3축 가속도 센서의 값, 평균, 표준편차, 수직 및 수평축의 데이터를 사용하였다. 분류 알고리즘은 Naive Bayes, Bayes Net, k-NN, SVM, Decision Tree, Multilayer perception, Logistic regression을 사용하였다. 연구결과 k-NN 알고리즘의 인식률이 98.7%로 다른 분류알고리즘에 비해 인식률이 높게 나타났다.

A Study on the Determination of the Risk-Loaded Premium using Risk Measures in the Credibility Theory (신뢰도이론에서 위험측도를 이용한 할증보험료 결정에 대한 고찰)

  • Kim, Hyun Tae;Jeon, Yongho
    • The Korean Journal of Applied Statistics
    • /
    • v.27 no.1
    • /
    • pp.71-87
    • /
    • 2014
  • The Bayes premium or the net premium in the credibility theory does not reflect the underlying tail risk. In this study we examine how the tail risk measures can be utilized in determining the risk premium. First, we show that the risk measures can not only provide the proper risk loading, but also allow the insurer to avoid the wrong decision made with the Bayesian premium alone. Second, it is illustrated that the rank of the tail thickness among different conditional loss distributions does not preserve for the corresponding predictive distributions, even if they share the identical prior variable. The implication of this result is that the risk loading for a contract should be based on the risk measure of the predictive loss distribution not the conditional one.

Enhancing Workers' Job Tenure Using Directions Derived from Data Mining Techniques (데이터 마이닝 기법을 활용한 근로자의 고용유지 강화 방안 개발)

  • An, Minuk;Kim, Taeun;Yoo, Donghee
    • The Journal of the Korea Contents Association
    • /
    • v.18 no.5
    • /
    • pp.265-279
    • /
    • 2018
  • This study conducted an experiment using data mining techniques to develop prediction models of worker job turnover. The experiment used data from the '2015 Graduate Occupational Mobility Survey' by the Korea Employment Information Service. We developed the prediction models using a decision tree, Bayes net, and artificial neural network. We found that the decision tree-based prediction model reported the best accuracy. We also found that the six influential factors affecting employees' turnover intention are type of working time, job status, full-time or not full-time, regular working hours per week, regular working days per week, and personal development opportunities. From the decision tree-based prediction model, we derived 12 rules of employee turnover for all job types. Using the derived rules, we proposed helpful directions for enhancing workers' job tenure. In addition, we analyzed the influential factors affecting employees' job turnover intention according to four job types and derived rules for each: office (ten rules), culture and art (nine rules), construction (four rules), and information technology (six rules). Using the derived rules, we proposed customized directions for improving the job tenure for each group.