• Title/Summary/Keyword: ROC 커브

Search Result 17, Processing Time 0.023 seconds

A Development of Wireless Sensor Networks for Collaborative Sensor Fusion Based Speaker Gender Classification (협동 센서 융합 기반 화자 성별 분류를 위한 무선 센서네트워크 개발)

  • Kwon, Ho-Min
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.12 no.2
    • /
    • pp.113-118
    • /
    • 2011
  • In this paper, we develop a speaker gender classification technique using collaborative sensor fusion for use in a wireless sensor network. The distributed sensor nodes remove the unwanted input data using the BER(Band Energy Ration) based voice activity detection, process only the relevant data, and transmit the hard labeled decisions to the fusion center where a global decision fusion is carried out. This takes advantages of power consumption and network resource management. The Bayesian sensor fusion and the global weighting decision fusion methods are proposed to achieve the gender classification. As the number of the sensor nodes varies, the Bayesian sensor fusion yields the best classification accuracy using the optimal operating points of the ROC(Receiver Operating Characteristic) curves_ For the weights used in the global decision fusion, the BER and MCL(Mutual Confidence Level) are employed to effectively combined at the fusion center. The simulation results show that as the number of the sensor nodes increases, the classification accuracy was even more improved in the low SNR(Signal to Noise Ration) condition.

Verification of the Suitability of Fine Dust and Air Quality Management Systems Based on Artificial Intelligence Evaluation Models

  • Heungsup Sim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.8
    • /
    • pp.165-170
    • /
    • 2024
  • This study aims to verify the accuracy of the air quality management system in Yangju City using an artificial intelligence (AI) evaluation model. The consistency and reliability of fine dust data were assessed by comparing public data from the Ministry of Environment with data from Yangju City's air quality management system. To this end, we analyzed the completeness, uniqueness, validity, consistency, accuracy, and integrity of the data. Exploratory statistical analysis was employed to compare data consistency. The results of the AI-based data quality index evaluation revealed no statistically significant differences between the two datasets. Among AI-based algorithms, the random forest model demonstrated the highest predictive accuracy, with its performance evaluated through ROC curves and AUC. Notably, the random forest model was identified as a valuable tool for optimizing the air quality management system. This study confirms that the reliability and suitability of fine dust data can be effectively assessed using AI-based model performance evaluation, contributing to the advancement of air quality management strategies.

Steganalysis Using Histogram Characteristic and Statistical Moments of Wavelet Subbands (웨이블릿 부대역의 히스토그램 특성과 통계적 모멘트를 이용한 스테그분석)

  • Hyun, Seung-Hwa;Park, Tae-Hee;Kim, Young-In;Kim, Yoo-Shin;Eom, Il-Kyu
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.47 no.6
    • /
    • pp.57-65
    • /
    • 2010
  • In this paper, we present a universal steganalysis scheme. The proposed method extract features of two types. First feature set is extracted from histogram characteristic of the wavelet subbands. Second feature set is determined by statistical moments of wavelet characteristic functions. 3-level wavelet decomposition is performed for stego image and cover image using the Haar wavelet basis. We extract one features from 9 high frequency subbands of 12 subbands. The number of second features is 39. We use total 48 features for steganalysis. Multi layer perceptron(MLP) is applied as classifier to distinguish between cover images and stego images. To evaluate the proposed steganalysis method, we use the CorelDraw image database. We test the performance of our proposed steganalysis method over LSB method, spread spectrum data hiding method, blind spread spectrum data hiding method and F5 data hiding method. The proposed method outperforms the previous methods in sensitivity, specificity, error rate and area under ROC curve, etc.

Score System for Operative Risk Evaluation in Coronary Artery Bypass Surgery (관상동맥 우회로술의 수술 위험인자에 대한 스코어 시스템)

  • Kang Joon-Kyu;Kim Chong-Wook;Sheen Seung-Soo;Chung Cheol-Hyun;Lee Jae-Won;Song Meong-Gun;Lee Jung-Sook;Song Hyun
    • Journal of Chest Surgery
    • /
    • v.39 no.10 s.267
    • /
    • pp.749-753
    • /
    • 2006
  • Background: The purpose of this study is to assess a score system for operative risk evaluation of CABG. Material and Method: From January 2001 to September 2005, retrospective study for various perioperative factors of 2993 cases was done. Result: The early operative mortality was 2.4% and the beta coefficients of 7 core variables related to it (preoperative LV dysfuction, preoperative renal failure, MI within 1 week, reoperation, combined surgery, preoperative atrial fibrillation, preoperative IABP) were adjusted to score system. ROC curve and Hosmer and Lemeshow goodness of fit test was done. Conclusion: This score system was effective in assessing operative risk of CABG. But It is necessary to gather larger volume of case and perform multicenter study.

A New Shock Index for Predicting Survival of Rats with Hemorrhagic Shock Using Perfusion and Lactate Concentration Ratio (흰쥐의 출혈성 쇼크에서 관류와 젖산 농도 비를 이용한 새로운 생존 예측 지표 개발)

  • Choi, Jae-Lim;Nam, Ki-Chang;Kwon, Min-Kyung;Jang, Kyung-Hwan;Kim, Deok-Won
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.48 no.4
    • /
    • pp.1-9
    • /
    • 2011
  • Hemorrhagic shock is a clinically widespread syndrome characterized by inadequate oxygenation and supply. It is important to diagnose hemorrhagic shock in its early stage for improving treatment effects and survival rate. However, an accurate diagnosis and treatment could be delayed in the early stage of hemorrhagic shock by evaluating only vital signs such as heart rate and blood pressure. There have been many studies for the early diagnosis of hemorrhagic shock, reporting that lactate concentration and perfusion were useful variables for tissue hypoxia and metabolic acidosis. In this study, we measured both perfusion using a laser Doppler flowmeter and lactate concentration from the volume controlled hemorrhagic shock using rats. We also proposed a new shock index which was calculated by dividing lactate concentration by perfusion for early diagnosis. As a result of the survival prediction by the proposed index with the receiver operating characteristic curve method, the sensitivity, specificity, and accuracy of survival were 90.0, 96.7 and 94.0%, respectively. The proposed index showed the fastest significant difference among the other parameters such as blood pressure and heart rate. It could offer early diagnosis and effective treatment for human hemorrhagic shock if it is applicable to humans.

Animal Infectious Diseases Prevention through Big Data and Deep Learning (빅데이터와 딥러닝을 활용한 동물 감염병 확산 차단)

  • Kim, Sung Hyun;Choi, Joon Ki;Kim, Jae Seok;Jang, Ah Reum;Lee, Jae Ho;Cha, Kyung Jin;Lee, Sang Won
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.137-154
    • /
    • 2018
  • Animal infectious diseases, such as avian influenza and foot and mouth disease, occur almost every year and cause huge economic and social damage to the country. In order to prevent this, the anti-quarantine authorities have tried various human and material endeavors, but the infectious diseases have continued to occur. Avian influenza is known to be developed in 1878 and it rose as a national issue due to its high lethality. Food and mouth disease is considered as most critical animal infectious disease internationally. In a nation where this disease has not been spread, food and mouth disease is recognized as economic disease or political disease because it restricts international trade by making it complex to import processed and non-processed live stock, and also quarantine is costly. In a society where whole nation is connected by zone of life, there is no way to prevent the spread of infectious disease fully. Hence, there is a need to be aware of occurrence of the disease and to take action before it is distributed. Epidemiological investigation on definite diagnosis target is implemented and measures are taken to prevent the spread of disease according to the investigation results, simultaneously with the confirmation of both human infectious disease and animal infectious disease. The foundation of epidemiological investigation is figuring out to where one has been, and whom he or she has met. In a data perspective, this can be defined as an action taken to predict the cause of disease outbreak, outbreak location, and future infection, by collecting and analyzing geographic data and relation data. Recently, an attempt has been made to develop a prediction model of infectious disease by using Big Data and deep learning technology, but there is no active research on model building studies and case reports. KT and the Ministry of Science and ICT have been carrying out big data projects since 2014 as part of national R &D projects to analyze and predict the route of livestock related vehicles. To prevent animal infectious diseases, the researchers first developed a prediction model based on a regression analysis using vehicle movement data. After that, more accurate prediction model was constructed using machine learning algorithms such as Logistic Regression, Lasso, Support Vector Machine and Random Forest. In particular, the prediction model for 2017 added the risk of diffusion to the facilities, and the performance of the model was improved by considering the hyper-parameters of the modeling in various ways. Confusion Matrix and ROC Curve show that the model constructed in 2017 is superior to the machine learning model. The difference between the2016 model and the 2017 model is that visiting information on facilities such as feed factory and slaughter house, and information on bird livestock, which was limited to chicken and duck but now expanded to goose and quail, has been used for analysis in the later model. In addition, an explanation of the results was added to help the authorities in making decisions and to establish a basis for persuading stakeholders in 2017. This study reports an animal infectious disease prevention system which is constructed on the basis of hazardous vehicle movement, farm and environment Big Data. The significance of this study is that it describes the evolution process of the prediction model using Big Data which is used in the field and the model is expected to be more complete if the form of viruses is put into consideration. This will contribute to data utilization and analysis model development in related field. In addition, we expect that the system constructed in this study will provide more preventive and effective prevention.

Bankruptcy prediction using an improved bagging ensemble (개선된 배깅 앙상블을 활용한 기업부도예측)

  • Min, Sung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.121-139
    • /
    • 2014
  • Predicting corporate failure has been an important topic in accounting and finance. The costs associated with bankruptcy are high, so the accuracy of bankruptcy prediction is greatly important for financial institutions. Lots of researchers have dealt with the topic associated with bankruptcy prediction in the past three decades. The current research attempts to use ensemble models for improving the performance of bankruptcy prediction. Ensemble classification is to combine individually trained classifiers in order to gain more accurate prediction than individual models. Ensemble techniques are shown to be very useful for improving the generalization ability of the classifier. Bagging is the most commonly used methods for constructing ensemble classifiers. In bagging, the different training data subsets are randomly drawn with replacement from the original training dataset. Base classifiers are trained on the different bootstrap samples. Instance selection is to select critical instances while deleting and removing irrelevant and harmful instances from the original set. Instance selection and bagging are quite well known in data mining. However, few studies have dealt with the integration of instance selection and bagging. This study proposes an improved bagging ensemble based on instance selection using genetic algorithms (GA) for improving the performance of SVM. GA is an efficient optimization procedure based on the theory of natural selection and evolution. GA uses the idea of survival of the fittest by progressively accepting better solutions to the problems. GA searches by maintaining a population of solutions from which better solutions are created rather than making incremental changes to a single solution to the problem. The initial solution population is generated randomly and evolves into the next generation by genetic operators such as selection, crossover and mutation. The solutions coded by strings are evaluated by the fitness function. The proposed model consists of two phases: GA based Instance Selection and Instance based Bagging. In the first phase, GA is used to select optimal instance subset that is used as input data of bagging model. In this study, the chromosome is encoded as a form of binary string for the instance subset. In this phase, the population size was set to 100 while maximum number of generations was set to 150. We set the crossover rate and mutation rate to 0.7 and 0.1 respectively. We used the prediction accuracy of model as the fitness function of GA. SVM model is trained on training data set using the selected instance subset. The prediction accuracy of SVM model over test data set is used as fitness value in order to avoid overfitting. In the second phase, we used the optimal instance subset selected in the first phase as input data of bagging model. We used SVM model as base classifier for bagging ensemble. The majority voting scheme was used as a combining method in this study. This study applies the proposed model to the bankruptcy prediction problem using a real data set from Korean companies. The research data used in this study contains 1832 externally non-audited firms which filed for bankruptcy (916 cases) and non-bankruptcy (916 cases). Financial ratios categorized as stability, profitability, growth, activity and cash flow were investigated through literature review and basic statistical methods and we selected 8 financial ratios as the final input variables. We separated the whole data into three subsets as training, test and validation data set. In this study, we compared the proposed model with several comparative models including the simple individual SVM model, the simple bagging model and the instance selection based SVM model. The McNemar tests were used to examine whether the proposed model significantly outperforms the other models. The experimental results show that the proposed model outperforms the other models.