• Title/Summary/Keyword: statistical analysis.

Search Result 17,816, Processing Time 0.041 seconds

A Computer Intensive Method for Modern Statistical Data Analysis I ; Bootststrap Method and Its Applications (통계적 데이터 분석방법을 위한 컴퓨터의 활용 I : 붓스트랩 이론과 응용+)

  • 전명식
    • The Korean Journal of Applied Statistics
    • /
    • v.3 no.1
    • /
    • pp.121-141
    • /
    • 1990
  • Computer intensive bootstrap methods are studied as a tool of statistics. Practical calculation and theoretical justification problem of the methods in estimating the sampling distribution and construction confidence region of parameters are discussed through several examples. Statistical meaning of the methods are also considered.

  • PDF

A Comparative Study of Microarray Data with Survival Times Based on Several Missing Mechanism

  • Kim Jee-Yun;Hwang Jin-Soo;Kim Seong-Sun
    • Communications for Statistical Applications and Methods
    • /
    • v.13 no.1
    • /
    • pp.101-111
    • /
    • 2006
  • One of the most widely used method of handling missingness in microarray data is the kNN(k Nearest Neighborhood) method. Recently Li and Gui (2004) suggested, so called PCR(Partial Cox Regression) method which deals with censored survival times and microarray data efficiently via kNN imputation method. In this article, we try to show that the way to treat missingness eventually affects the further statistical analysis.

A Study on Distribution Based on the Normalized Sample Lorenz Curve

  • Suk-Bok kang;Cho, Young-Suk
    • Communications for Statistical Applications and Methods
    • /
    • v.8 no.1
    • /
    • pp.185-192
    • /
    • 2001
  • Using the Lorenz curve that is proved to be a powerful tool to measure the income inequality within a population of income receivers, we propose the normalized sample Lorenz curve for the goodness-of-fit test that is very important test in statistical analysis. For two hodgkin's disease data sets, we compare the Q-Q plot and the proposed normalized sample Lorenz curve.

  • PDF

Tree-structured Classification based on Variable Splitting

  • Ahn, Sung-Jin
    • Communications for Statistical Applications and Methods
    • /
    • v.2 no.1
    • /
    • pp.74-88
    • /
    • 1995
  • This article introduces a unified method of choosing the most explanatory and significant multiway partitions for classification tree design and analysis. The method is derived on the impurity reduction (IR) measure of divergence, which is proposed to extend the proportional-reduction-in-error (PRE) measure in the decision-theory context. For the method derivation, the IR measure is analyzed to characterize its statistical properties which are used to consistently handle the subjects of feature formation, feature selection, and feature deletion required in the associated classification tree construction. A numerical example is considered to illustrate the proposed approach.

  • PDF

Understanding and Misuse Type of Quality Improvement Tools According to the Kind of Data and the Number of Population in DMAIC Process of Six Sigma (식스시그마 DMAIC 프로세스에서 모집단의 수와 데이터 종류에 따른 품질개선 기법의 오적용 유형 및 이해)

  • Choi, Sung-Woon
    • Proceedings of the Safety Management and Science Conference
    • /
    • 2010.04a
    • /
    • pp.509-517
    • /
    • 2010
  • The paper proposes the misuse types of statistical quality tools according to the kind of data and the number of population in DMAIC process of six sigma. The result presented in this paper can be extended to the QC story 15 steps of QC circle. The study also provides the improvement methods about control chart, measurement system analysis, statistical difference, and practical equivalence.

  • PDF

Development of Web-based Quality & Reliability System for Bootstrap on the Internet Environment (인터넷 환경에서 붓스트랩 품질 및 신뢰성 시스템의 개발)

  • Choi Sung woon;Lim In sup
    • Journal of the Korea Safety Management & Science
    • /
    • v.7 no.1
    • /
    • pp.147-157
    • /
    • 2005
  • Recently, growth of internet causes rapid changes in many areas of statistics such as statistical computation and analysis. Especially, bootstrap is the most interesting statistical methods applying computer resampling simulation. In this paper, we try to present how to use a method of bootstrap on the internet. We also develop to user a statistical system which is programed with ASP for user to handle easily in manufacturing system.

Test of Normality Based on the Transformed Lorenz Curve

  • Kang, Suk-Bok;Cho, Young-Suk
    • Communications for Statistical Applications and Methods
    • /
    • v.6 no.3
    • /
    • pp.901-908
    • /
    • 1999
  • Using the Transformed Lorenz curve which is introduced by Cho et al.(1999) we propose the test statistic for testing of normality that is very important test in statistical analysis and compare the proposed test statistic with the Shapiro and Wilk's W test statistic in terms of the power of test through by Monte Carlo method.

  • PDF

Resistant Principal Factor Analysis

  • Park, Youg-Seok;Byun, Ho-Seon
    • Journal of the Korean Statistical Society
    • /
    • v.25 no.1
    • /
    • pp.67-80
    • /
    • 1996
  • Factor analysis is a multivariate technique for describing the in-terrelationship among many variables in terms of a few underlying but unobservable random variables called factors. There are various approaches for this factor analysis. In particular, principal factor analysis is one of the most popular methods. This follows the mathematical algorithm of the principal component analysis based on the singular value decomposition. But it is known that the singular value decomposition is not resistant, i.e., it is very sensitive to small changes in the input data. In this article, using the resistant singular value decomposition of Choi and Huh (1994), we derive a resistant principal factor analysis relatively little influenced by notable observations.

  • PDF

Exploring COVID-19 in mainland China during the lockdown of Wuhan via functional data analysis

  • Li, Xing;Zhang, Panpan;Feng, Qunqiang
    • Communications for Statistical Applications and Methods
    • /
    • v.29 no.1
    • /
    • pp.103-125
    • /
    • 2022
  • In this paper, we analyze the time series data of the case and death counts of COVID-19 that broke out in China in December, 2019. The study period is during the lockdown of Wuhan. We exploit functional data analysis methods to analyze the collected time series data. The analysis is divided into three parts. First, the functional principal component analysis is conducted to investigate the modes of variation. Second, we carry out the functional canonical correlation analysis to explore the relationship between confirmed and death cases. Finally, we utilize a clustering method based on the Expectation-Maximization (EM) algorithm to run the cluster analysis on the counts of confirmed cases, where the number of clusters is determined via a cross-validation approach. Besides, we compare the clustering results with some migration data available to the public.