• Title/Summary/Keyword: Statistical Modeling

Search Result 1,200, Processing Time 0.025 seconds

Statistical Modeling of Pretilt Angle Control using Ion-beam Alignment on Nitrogen Doped Diamond-like Carbon Thin Film

  • Kang, Hee-Jin;Lee, Jung-Hwan;Han, Jung-Min;Yun, Il-Gu;Seo, Dae-Shik
    • Transactions on Electrical and Electronic Materials
    • /
    • v.7 no.6
    • /
    • pp.297-300
    • /
    • 2006
  • The response surface modeling of the pretilt angle control using ion-beam (IB) alignment on nitrogen doped diamond-like carbon (NDLC) thin film layer is investigated. This modeling is used to analyze the variation of the pretilt angle under various process conditions. IB exposure angle and IB exposure time are considered as input factors. The analysis of variance technique is settled to analyze the statistical significance, and effect plots are also investigated to examine the relationships between the process parameters and the response. The model can allow us to reliably predict the pretilt angle with respect to the varying process conditions.

A Flexible Modeling Approach for Current Status Survival Data via Pseudo-Observations

  • Han, Seungbong;Andrei, Adin-Cristian;Tsui, Kam-Wah
    • The Korean Journal of Applied Statistics
    • /
    • v.25 no.6
    • /
    • pp.947-958
    • /
    • 2012
  • When modeling event times in biomedical studies, the outcome might be incompletely observed. In this paper, we assume that the outcome is recorded as current status failure time data. Despite well-developed literature the routine practical use of many current status data modeling methods remains infrequent due to the lack of specialized statistical software, the difficulty to assess model goodness-of-fit, as well as the possible loss of information caused by covariate grouping or discretization. We propose a model based on pseudo-observations that is convenient to implement and that allows for flexibility in the choice of the outcome. Parameter estimates are obtained based on generalized estimating equations. Examples from studies in bile duct hyperplasia and breast cancer in conjunction with simulated data illustrate the practical advantages of this model.

A rolling analysis on the prediction of value at risk with multivariate GARCH and copula

  • Bai, Yang;Dang, Yibo;Park, Cheolwoo;Lee, Taewook
    • Communications for Statistical Applications and Methods
    • /
    • v.25 no.6
    • /
    • pp.605-618
    • /
    • 2018
  • Risk management has been a crucial part of the daily operations of the financial industry over the past two decades. Value at Risk (VaR), a quantitative measure introduced by JP Morgan in 1995, is the most popular and simplest quantitative measure of risk. VaR has been widely applied to the risk evaluation over all types of financial activities, including portfolio management and asset allocation. This paper uses the implementations of multivariate GARCH models and copula methods to illustrate the performance of a one-day-ahead VaR prediction modeling process for high-dimensional portfolios. Many factors, such as the interaction among included assets, are included in the modeling process. Additionally, empirical data analyses and backtesting results are demonstrated through a rolling analysis, which help capture the instability of parameter estimates. We find that our way of modeling is relatively robust and flexible.

A Bayesian joint model for continuous and zero-inflated count data in developmental toxicity studies

  • Hwang, Beom Seuk
    • Communications for Statistical Applications and Methods
    • /
    • v.29 no.2
    • /
    • pp.239-250
    • /
    • 2022
  • In many applications, we frequently encounter correlated multiple outcomes measured on the same subject. Joint modeling of such multiple outcomes can improve efficiency of inference compared to independent modeling. For instance, in developmental toxicity studies, fetal weight and number of malformed pups are measured on the pregnant dams exposed to different levels of a toxic substance, in which the association between such outcomes should be taken into account in the model. The number of malformations may possibly have many zeros, which should be analyzed via zero-inflated count models. Motivated by applications in developmental toxicity studies, we propose a Bayesian joint modeling framework for continuous and count outcomes with excess zeros. In our model, zero-inflated Poisson (ZIP) regression model would be used to describe count data, and a subject-specific random effects would account for the correlation across the two outcomes. We implement a Bayesian approach using MCMC procedure with data augmentation method and adaptive rejection sampling. We apply our proposed model to dose-response analysis in a developmental toxicity study to estimate the benchmark dose in a risk assessment.

A review of tree-based Bayesian methods

  • Linero, Antonio R.
    • Communications for Statistical Applications and Methods
    • /
    • v.24 no.6
    • /
    • pp.543-559
    • /
    • 2017
  • Tree-based regression and classification ensembles form a standard part of the data-science toolkit. Many commonly used methods take an algorithmic view, proposing greedy methods for constructing decision trees; examples include the classification and regression trees algorithm, boosted decision trees, and random forests. Recent history has seen a surge of interest in Bayesian techniques for constructing decision tree ensembles, with these methods frequently outperforming their algorithmic counterparts. The goal of this article is to survey the landscape surrounding Bayesian decision tree methods, and to discuss recent modeling and computational developments. We provide connections between Bayesian tree-based methods and existing machine learning techniques, and outline several recent theoretical developments establishing frequentist consistency and rates of convergence for the posterior distribution. The methodology we present is applicable for a wide variety of statistical tasks including regression, classification, modeling of count data, and many others. We illustrate the methodology on both simulated and real datasets.

A New Methodology for Software Reliability based on Statistical Modeling

  • Avinash S;Y.Srinivas;P.Annan naidu
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.9
    • /
    • pp.157-161
    • /
    • 2023
  • Reliability is one of the computable quality features of the software. To assess the reliability the software reliability growth models(SRGMS) are used at different test times based on statistical learning models. In all situations, Tradational time-based SRGMS may not be enough, and such models cannot recognize errors in small and medium sized applications.Numerous traditional reliability measures are used to test software errors during application development and testing. In the software testing and maintenance phase, however, new errors are taken into consideration in real time in order to decide the reliability estimate. In this article, we suggest using the Weibull model as a computational approach to eradicate the problem of software reliability modeling. In the suggested model, a new distribution model is suggested to improve the reliability estimation method. We compute the model developed and stabilize its efficiency with other popular software reliability growth models from the research publication. Our assessment results show that the proposed Model is worthier to S-shaped Yamada, Generalized Poisson, NHPP.

A Study on Statistical Parameters for the Evaluation of Regional Air Quality Modeling Results - Focused on Fine Dust Modeling - (지역규모 대기질 모델 결과 평가를 위한 통계 검증지표 활용 - 미세먼지 모델링을 중심으로 -)

  • Kim, Cheol-Hee;Lee, Sang-Hyun;Jang, Min;Chun, Sungnam;Kang, Suji;Ko, Kwang-Kun;Lee, Jong-Jae;Lee, Hyo-Jung
    • Journal of Environmental Impact Assessment
    • /
    • v.29 no.4
    • /
    • pp.272-285
    • /
    • 2020
  • We investigated statistical evaluation parameters for 3D meteorological and air quality models and selected several quantitative indicator references, and summarized the reference values of the statistical parameters for domestic air quality modeling researcher. The finally selected 9 statistical parameters are MB (Mean Bias), ME (Mean Error), MNB (Mean Normalized Bias Error), MNE (Mean Absolute Gross Error), RMSE (Root Mean Square Error), IOA (Index of Agreement), R (Correlation Coefficient), FE (Fractional Error), FB (Fractional Bias), and the associated reference values are summarized. The results showed that MB and ME have been widely used in evaluating the meteorological model output, and NMB and NME are most frequently used for air quality model results. In addition, discussed are the presentation diagrams such as Soccer Plot, Taylor diagram, and Q-Q (Quantile-Quantile) diagram. The current results from our study is expected to be effectively used as the statistical evaluation parameters suitable for situation in Korea considering various characteristics such as including the mountainous surface areas.

Modeling Hydrogen Peroxide Bleaching Process to Predict Optical Properties of a Bleached CMP Pulp

  • Hatam Abouzar;Pourtahmasi Kambiz;Resalati Hossein;Lohrasebi A. Hossein
    • Proceedings of the Korea Technical Association of the Pulp and Paper Industry Conference
    • /
    • 2006.06b
    • /
    • pp.365-372
    • /
    • 2006
  • In this paper, the possibility of statistical modeling from the pulp and peroxide bleaching condition variables to predict optical properties of a bleached chemimechanical pulp used in a newsprint paper machine at Mazandaran Wood and Paper Industries Company (MWPI) was studied. Due to the variations in the opacity and the brightness of the bleached pulp at MWPI and to tackle this problem, it was decided to study the possibility of modeling the bleaching process. To achieve this purpose, Multi-variate Regression Analysis was used for model building and it was found that there is a relationship between independent variables and pulp brightness as well as pulp opacity, consequently, two models were constructed. Then, model validation was carried out using new data set in the bleaching plant at MWPI to test model predictive ability and its performance.

  • PDF

Review of Statistical Methods for Evaluating the Performance of Survival or Other Time-to-Event Prediction Models (from Conventional to Deep Learning Approaches)

  • Seo Young Park;Ji Eun Park;Hyungjin Kim;Seong Ho Park
    • Korean Journal of Radiology
    • /
    • v.22 no.10
    • /
    • pp.1697-1707
    • /
    • 2021
  • The recent introduction of various high-dimensional modeling methods, such as radiomics and deep learning, has created a much greater diversity in modeling approaches for survival prediction (or, more generally, time-to-event prediction). The newness of the recent modeling approaches and unfamiliarity with the model outputs may confuse some researchers and practitioners about the evaluation of the performance of such models. Methodological literacy to critically appraise the performance evaluation of the models and, ideally, the ability to conduct such an evaluation would be needed for those who want to develop models or apply them in practice. This article intends to provide intuitive, conceptual, and practical explanations of the statistical methods for evaluating the performance of survival prediction models with minimal usage of mathematical descriptions. It covers from conventional to deep learning methods, and emphasis has been placed on recent modeling approaches. This review article includes straightforward explanations of C indices (Harrell's C index, etc.), time-dependent receiver operating characteristic curve analysis, calibration plot, other methods for evaluating the calibration performance, and Brier score.

Modeling Extreme Values of Ground-Level Ozone Based on Threshold Methods for Markov Chains

  • Seokhoon Yun
    • Communications for Statistical Applications and Methods
    • /
    • v.3 no.2
    • /
    • pp.249-273
    • /
    • 1996
  • This paper reviews and develops several statistical models for extreme values, based on threshold methodology. Extreme values of a time series are modeled in terms of tails which are defined as truncated forms of original variables, and Markov property is imposed on the tails. Tails of the generalized extreme value distribution and a multivariate extreme value distributively, of the tails of the series. These models are then applied to real ozone data series collected in the Chicago area. A major concern is given to detecting any possible trend in the extreme values.

  • PDF