• Title/Summary/Keyword: predictive accuracy

Search Result 797, Processing Time 0.023 seconds

Diagnostic Accuracy of Percutaneous Transthoracic Needle Lung Biopsies: A Multicenter Study

  • Kyung Hee Lee;Kun Young Lim;Young Joo Suh;Jin Hur;Dae Hee Han;Mi-Jin Kang;Ji Yung Choo;Cherry Kim;Jung Im Kim;Soon Ho Yoon;Woojoo Lee;Chang Min Park
    • Korean Journal of Radiology
    • /
    • v.20 no.8
    • /
    • pp.1300-1310
    • /
    • 2019
  • Objective: To measure the diagnostic accuracy of percutaneous transthoracic needle lung biopsies (PTNBs) on the basis of the intention-to-diagnose principle and identify risk factors for diagnostic failure of PTNBs in a multi-institutional setting. Materials and Methods: A total of 9384 initial PTNBs performed in 9239 patients (mean patient age, 65 years [range, 20-99 years]) from January 2010 to December 2014 were included. The accuracy, sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) of PTNBs for diagnosis of malignancy were measured. The proportion of diagnostic failures was measured, and their risk factors were identified. Results: The overall accuracy, sensitivity, specificity, PPV, and NPV were 91.1% (95% confidence interval [CI], 90.6-91.7%), 92.5% (95% CI, 91.9-93.1%), 86.5% (95% CI, 85.0-87.9%), 99.2% (95% CI, 99.0-99.4%), and 84.3% (95% CI, 82.7-85.8%), respectively. The proportion of diagnostic failures was 8.9% (831 of 9384; 95% CI, 8.3-9.4%). The independent risk factors for diagnostic failures were lesions ≤ 1 cm in size (adjusted odds ratio [AOR], 1.86; 95% CI, 1.23-2.81), lesion size 1.1-2 cm (1.75; 1.45-2.11), subsolid lesions (1.81; 1.32-2.49), use of fine needle aspiration only (2.43; 1.80-3.28), final diagnosis of benign lesions (2.18; 1.84-2.58), and final diagnosis of lymphomas (10.66; 6.21-18.30). Use of cone-beam CT (AOR, 0.31; 95% CI, 0.13-0.75) and conventional CT-guidance (0.55; 0.32-0.94) reduced diagnostic failures. Conclusion: The accuracy of PTNB for diagnosis of malignancy was fairly high in our large-scale multi-institutional cohort. The identified risk factors for diagnostic failure may help reduce diagnostic failure and interpret the biopsy results.

The Prediction of Export Credit Guarantee Accident using Machine Learning (기계학습을 이용한 수출신용보증 사고예측)

  • Cho, Jaeyoung;Joo, Jihwan;Han, Ingoo
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.83-102
    • /
    • 2021
  • The government recently announced various policies for developing big-data and artificial intelligence fields to provide a great opportunity to the public with respect to disclosure of high-quality data within public institutions. KSURE(Korea Trade Insurance Corporation) is a major public institution for financial policy in Korea, and thus the company is strongly committed to backing export companies with various systems. Nevertheless, there are still fewer cases of realized business model based on big-data analyses. In this situation, this paper aims to develop a new business model which can be applied to an ex-ante prediction for the likelihood of the insurance accident of credit guarantee. We utilize internal data from KSURE which supports export companies in Korea and apply machine learning models. Then, we conduct performance comparison among the predictive models including Logistic Regression, Random Forest, XGBoost, LightGBM, and DNN(Deep Neural Network). For decades, many researchers have tried to find better models which can help to predict bankruptcy since the ex-ante prediction is crucial for corporate managers, investors, creditors, and other stakeholders. The development of the prediction for financial distress or bankruptcy was originated from Smith(1930), Fitzpatrick(1932), or Merwin(1942). One of the most famous models is the Altman's Z-score model(Altman, 1968) which was based on the multiple discriminant analysis. This model is widely used in both research and practice by this time. The author suggests the score model that utilizes five key financial ratios to predict the probability of bankruptcy in the next two years. Ohlson(1980) introduces logit model to complement some limitations of previous models. Furthermore, Elmer and Borowski(1988) develop and examine a rule-based, automated system which conducts the financial analysis of savings and loans. Since the 1980s, researchers in Korea have started to examine analyses on the prediction of financial distress or bankruptcy. Kim(1987) analyzes financial ratios and develops the prediction model. Also, Han et al.(1995, 1996, 1997, 2003, 2005, 2006) construct the prediction model using various techniques including artificial neural network. Yang(1996) introduces multiple discriminant analysis and logit model. Besides, Kim and Kim(2001) utilize artificial neural network techniques for ex-ante prediction of insolvent enterprises. After that, many scholars have been trying to predict financial distress or bankruptcy more precisely based on diverse models such as Random Forest or SVM. One major distinction of our research from the previous research is that we focus on examining the predicted probability of default for each sample case, not only on investigating the classification accuracy of each model for the entire sample. Most predictive models in this paper show that the level of the accuracy of classification is about 70% based on the entire sample. To be specific, LightGBM model shows the highest accuracy of 71.1% and Logit model indicates the lowest accuracy of 69%. However, we confirm that there are open to multiple interpretations. In the context of the business, we have to put more emphasis on efforts to minimize type 2 error which causes more harmful operating losses for the guaranty company. Thus, we also compare the classification accuracy by splitting predicted probability of the default into ten equal intervals. When we examine the classification accuracy for each interval, Logit model has the highest accuracy of 100% for 0~10% of the predicted probability of the default, however, Logit model has a relatively lower accuracy of 61.5% for 90~100% of the predicted probability of the default. On the other hand, Random Forest, XGBoost, LightGBM, and DNN indicate more desirable results since they indicate a higher level of accuracy for both 0~10% and 90~100% of the predicted probability of the default but have a lower level of accuracy around 50% of the predicted probability of the default. When it comes to the distribution of samples for each predicted probability of the default, both LightGBM and XGBoost models have a relatively large number of samples for both 0~10% and 90~100% of the predicted probability of the default. Although Random Forest model has an advantage with regard to the perspective of classification accuracy with small number of cases, LightGBM or XGBoost could become a more desirable model since they classify large number of cases into the two extreme intervals of the predicted probability of the default, even allowing for their relatively low classification accuracy. Considering the importance of type 2 error and total prediction accuracy, XGBoost and DNN show superior performance. Next, Random Forest and LightGBM show good results, but logistic regression shows the worst performance. However, each predictive model has a comparative advantage in terms of various evaluation standards. For instance, Random Forest model shows almost 100% accuracy for samples which are expected to have a high level of the probability of default. Collectively, we can construct more comprehensive ensemble models which contain multiple classification machine learning models and conduct majority voting for maximizing its overall performance.

A Framework for Managing Approximation Models in place of Expensive Simulations in Optimization (최적화에서의 근사모델 관리기법의 활용)

  • 양영순;장범선;연윤석
    • Proceedings of the Computational Structural Engineering Institute Conference
    • /
    • 2000.04b
    • /
    • pp.159-167
    • /
    • 2000
  • In optimization problems, computationally intensive or expensive simulations hinder the use of standard optimization techniques because the computational expense is too heavy to implement them at each iteration of the optimization algorithm. Therefore, those expensive simulations are often replaced with approximation models which can be evaluated nearly free. However, because of the limited accuracy of the approximation models, it is practically impossible to find an exact optimal point of the original problem. Significant efforts have been made to overcome this problem. The approximation models are sequentially updated during the iterative optimization process such that interesting design points are included. The interesting points have a strong influence on making the approximation model capture an overall trend of the original function or improving the accuracy of the approximation in the vicinity of a minimizer. They are successively determined at each iteration by utilizing the predictive ability of the approximation model. This paper will focuses on those approaches and introduces various approximation methods.

  • PDF

A Highly Efficient Aeroelastic Optimization Method Based on a Surrogate Model

  • Zhiqiang, Wan;Xiaozhe, Wang;Chao, Yang
    • International Journal of Aeronautical and Space Sciences
    • /
    • v.17 no.4
    • /
    • pp.491-500
    • /
    • 2016
  • This paper presents a highly efficient aeroelastic optimization method based on a surrogate model; the model is verified by considering the case of a high-aspect-ratio composite wing. Optimization frameworks using the Kriging model and genetic algorithm (GA), the Kriging model and improved particle swarm optimization (IPSO), and the back propagation neural network model (BP) and IPSO are presented. The feasibility of the method is verified, as the model can improve the optimization efficiency while also satisfying the engineering requirements. Moreover, the effects of the number of design variables and number of constraints on the optimization efficiency and objective function are analysed in detail. The accuracy of two surrogate models in aeroelastic optimization is also compared. The Kriging model is constructed more conveniently, and its predictive accuracy of the aeroelastic responses also satisfies the engineering requirements. According to the case of a high-aspect-ratio composite wing, the GA is better at global optimization.

Accurate Current Reference Generator for Active Power Filters (능동전력필터의 정밀 기준신호 발생기)

  • Bae Byung-Yeol;Jon Young-Soo;Han Byung-Moon;Soh Yong-Choel
    • Proceedings of the KIPE Conference
    • /
    • 2004.07b
    • /
    • pp.575-578
    • /
    • 2004
  • The performance of an active power filter(APF) depends on the inverter characteristic, the control method, and the accuracy of reference signal generator. The accuracy of reference generator is the most critical item to determine the performance of active power filter. This paper introduces a novel reference signal generator composed of improved adaptive predictive filter. The performance of proposed reference signal generator was first verified through a simulation with MATLAB. Furthermore, the application of feasibility was evaluated through experimenting with a single-phase APF prototype based on the proposed reference generator, which was implemented using the TMS320C31 floating-point signal processor. Both simulations and experimental results confirm that our reference signal generator can be used successfully in practical active power filters.

  • PDF

Maintenance-based prognostics of nuclear plant equipment for long-term operation

  • Welz, Zachary;Coble, Jamie;Upadhyaya, Belle;Hines, Wes
    • Nuclear Engineering and Technology
    • /
    • v.49 no.5
    • /
    • pp.914-919
    • /
    • 2017
  • While industry understands the importance of keeping equipment operational and well maintained, the importance of tracking maintenance information in reliability models is often overlooked. Prognostic models can be used to predict the failure times of critical equipment, but more often than not, these models assume that all maintenance actions are the same or do not consider maintenance at all. This study investigates the influence of integrating maintenance information on prognostic model prediction accuracy. By incorporating maintenance information to develop maintenance-dependent prognostic models, prediction accuracy was improved by more than 40% compared with traditional maintenance-independent models. This study acts as a proof of concept, showing the importance of utilizing maintenance information in modern prognostics for industrial equipment.

Enhance Health Risks Prediction Mechanism in the Cloud Using RT-TKRIBC Technique

  • Konduru, Venkateswara Raju;Bharamgoudra, Manjula R
    • Journal of information and communication convergence engineering
    • /
    • v.19 no.3
    • /
    • pp.166-174
    • /
    • 2021
  • A large volume of patient data is generated from various devices used in healthcare applications. With increase in the volume of data generated in the healthcare industry, more wellness monitoring is required. A cloud-enabled analysis of healthcare data that predicts patient risk factors is required. Machine learning techniques have been developed to address these medical care problems. A novel technique called the radix-trie-based Tanimoto kernel regressive infomax boost classification (RT-TKRIBC) technique is introduced to analyze the heterogeneous health data in the cloud to predict the health risks and send alerts. The infomax boost ensemble technique improves the prediction accuracy by finding the maximum mutual information, thereby minimizing the mean square error. The performance evaluation of the proposed RT-TKRIBC technique is realized through extensive simulations in the cloud environment, which provides better prediction accuracy and less prediction time than those provided by the state-of-the-art methods.

Real-time implementation and performance evaluation of speech classifiers in speech analysis-synthesis

  • Kumar, Sandeep
    • ETRI Journal
    • /
    • v.43 no.1
    • /
    • pp.82-94
    • /
    • 2021
  • In this work, six voiced/unvoiced speech classifiers based on the autocorrelation function (ACF), average magnitude difference function (AMDF), cepstrum, weighted ACF (WACF), zero crossing rate and energy of the signal (ZCR-E), and neural networks (NNs) have been simulated and implemented in real time using the TMS320C6713 DSP starter kit. These speech classifiers have been integrated into a linear-predictive-coding-based speech analysis-synthesis system and their performance has been compared in terms of the percentage of the voiced/unvoiced classification accuracy, speech quality, and computation time. The results of the percentage of the voiced/unvoiced classification accuracy and speech quality show that the NN-based speech classifier performs better than the ACF-, AMDF-, cepstrum-, WACF- and ZCR-E-based speech classifiers for both clean and noisy environments. The computation time results show that the AMDF-based speech classifier is computationally simple, and thus its computation time is less than that of other speech classifiers, while that of the NN-based speech classifier is greater compared with other classifiers.

A Generation and Accuracy Evaluation of Common Metadata Prediction Model Using Public Bicycle Data and Imputation Method

  • Kim, Jong-Chan;Jung, Se-Hoon
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.2
    • /
    • pp.287-296
    • /
    • 2022
  • Today, air pollution is becoming a severe issue worldwide and various policies are being implemented to solve environmental pollution. In major cities, public bicycles are installed and operated to reduce pollution and solve transportation problems, and operational information is collected in real time. However, research using public bicycle operation information data has not been processed. This study uses the daily weather data of Korea Meteorological Agency and real-time air pollution data of Korea Environment Corporation to predict the amount of daily rental bicycles. Cross- validation, principal component analysis and multiple regression analysis were used to determine the independent variables of the predictive model. Then, the study selected the elements that satisfy the significance level, constructed a model, predicted the amount of daily rental bicycles, and measured the accuracy.

Cross-Project Pooling of Defects for Handling Class Imbalance

  • Catherine, J.M.;Djodilatchoumy, S
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.10
    • /
    • pp.11-16
    • /
    • 2022
  • Applying predictive analytics to predict software defects has improved the overall quality and decreased maintenance costs. Many supervised and unsupervised learning algorithms have been used for defect prediction on publicly available datasets. Most of these datasets suffer from an imbalance in the output classes. We study the impact of class imbalance in the defect datasets on the efficiency of the defect prediction model and propose a CPP method for handling imbalances in the dataset. The performance of the methods is evaluated using measures like Matthew's Correlation Coefficient (MCC), Recall, and Accuracy measures. The proposed sampling technique shows significant improvement in the efficiency of the classifier in predicting defects.