• 제목/요약/키워드: Classical Statistical Method

검색결과 109건 처리시간 0.026초

Improved Statistical Testing of Two-class Microarrays with a Robust Statistical Approach

  • Oh, Hee-Seok;Jang, Dong-Ik;Oh, Seung-Yoon;Kim, Hee-Bal
    • Interdisciplinary Bio Central
    • /
    • 제2권2호
    • /
    • pp.4.1-4.6
    • /
    • 2010
  • The most common type of microarray experiment has a simple design using microarray data obtained from two different groups or conditions. A typical method to identify differentially expressed genes (DEGs) between two conditions is the conventional Student's t-test. The t-test is based on the simple estimation of the population variance for a gene using the sample variance of its expression levels. Although empirical Bayes approach improves on the t-statistic by not giving a high rank to genes only because they have a small sample variance, the basic assumption for this is same as the ordinary t-test which is the equality of variances across experimental groups. The t-test and empirical Bayes approach suffer from low statistical power because of the assumption of normal and unimodal distributions for the microarray data analysis. We propose a method to address these problems that is robust to outliers or skewed data, while maintaining the advantages of the classical t-test or modified t-statistics. The resulting data transformation to fit the normality assumption increases the statistical power for identifying DEGs using these statistics.

Numerical and statistical analysis of permeability of concrete as a random heterogeneous composite

  • Zhou, Chunsheng;Li, Kefei
    • Computers and Concrete
    • /
    • 제7권5호
    • /
    • pp.469-482
    • /
    • 2010
  • This paper investigates the concrete permeability through a numerical and statistical approach. Concrete is considered as a random heterogeneous composite of three phases: aggregates, interfacial transition zones (ITZ) and matrix. The paper begins with some classical bound and estimate theories applied to concrete permeability and the influence of ITZ on these bound and estimate values is discussed. Numerical samples for permeability analysis are established through random aggregate structure (RAS) scheme, each numerical sample containing randomly distributed aggregates coated with ITZ and dispersed in a homogeneous matrix. The volumetric fraction of aggregates is fixed and the size distribution of aggregates observes Fuller's curve. Then finite element method is used to solve the steady permeation problem on 2D numerical samples and the overall permeability is deduced from flux-pressure relation. The impact of ITZ on overall permeability is analyzed in terms of ITZ width and contrast ratio between ITZ and matrix permeabilities. Hereafter, 3680 samples are generated for 23 sample sizes and 4 contrast ratios, and statistical analysis is performed on the permeability dispersion in terms of sample size and ITZ characteristics. By sample theory, the size of representative volume element (RVE) for permeability is then quantified considering sample realization number and expected error. Concluding remarks are provided for the impact of ITZ on concrete permeability and its statistical characteristics.

Tests for homogeneity of proportions in clustered binomial data

  • Jeong, Kwang Mo
    • Communications for Statistical Applications and Methods
    • /
    • 제23권5호
    • /
    • pp.433-444
    • /
    • 2016
  • When we observe binary responses in a cluster (such as rat lab-subjects), they are usually correlated to each other. In clustered binomial counts, the independence assumption is violated and we encounter an extra-variation. In the presence of extra-variation, the ordinary statistical analyses of binomial data are inappropriate to apply. In testing the homogeneity of proportions between several treatment groups, the classical Pearson chi-squared test has a severe flaw in the control of Type I error rates. We focus on modifying the chi-squared statistic by incorporating variance inflation factors. We suggest a method to adjust data in terms of dispersion estimate based on a quasi-likelihood model. We explain the testing procedure via an illustrative example as well as compare the performance of a modified chi-squared test with competitive statistics through a Monte Carlo study.

SPC 차트를 이용한 포트폴리오 관리 (Portfolio Management Using Statistical Process Control Chart)

  • 김동섭;류홍서
    • 산업공학
    • /
    • 제20권2호
    • /
    • pp.94-102
    • /
    • 2007
  • Portfolio management deals with decision making on 'when' and 'how' to revise an existing portfolio. In this paper, we show that a classical statistical process control (SPC) chart for normal data, a wellestablished tool in quality engineering, can effectively be used for signaling times for revising a portfolio. Noting that the day-to-day performance of a portfolio may be auto-correlated, we use the exponentially weighted moving average center-line chart to develop an automatic portfolio management procedure. The portfolio management procedure is extensively tested on historical data of equities traded in the Korea Exchange (KRX), the American Stock Exchange (AMEX), and the New York Stock Exchange (NYSE). In comparison with the performances of the KOSPI, XAX, and NYA indices during the same time periods, results from these experiments show that SPC chart-based portfolio revision presents itself a convenient and reliable method for optimally managing portfolios.

A new extension of Lindley distribution: modified validation test, characterizations and different methods of estimation

  • Ibrahim, Mohamed;Yadav, Abhimanyu Singh;Yousof, Haitham M.;Goual, Hafida;Hamedani, G.G.
    • Communications for Statistical Applications and Methods
    • /
    • 제26권5호
    • /
    • pp.473-495
    • /
    • 2019
  • In this paper, a new extension of Lindley distribution has been introduced. Certain characterizations based on truncated moments, hazard and reverse hazard function, conditional expectation of the proposed distribution are presented. Besides, these characterizations, other statistical/mathematical properties of the proposed model are also discussed. The estimation of the parameters is performed through different classical methods of estimation. Bayes estimation is computed under gamma informative prior under the squared error loss function. The performances of all estimation methods are studied via Monte Carlo simulations in mean square error sense. The potential of the proposed model is analyzed through two data sets. A modified goodness-of-fit test using the Nikulin-Rao-Robson statistic test is investigated via two examples and is observed that the new extension might be used as an alternative lifetime model.

이상점 탐지를 위한 일반화 우도비 검정 (A Generalized Likelihood Ratio Test in Outlier Detection)

  • Jang Sun Baek
    • 응용통계연구
    • /
    • 제7권2호
    • /
    • pp.225-237
    • /
    • 1994
  • 본 연구에서는 핵확산 감시와 관련된 이상점 탐지를 위한 일반화 우도비 검정 방법이 개발되었다. 고전적인 이상점 탐지방법들이 연속형 변수만을 고려한 반면, 본 연구에서 제안된 방법은 연속형 변수, 이산형 변수, 혹은 이산형과 연속형이 혼합된 변수들에 모두 적용될 수 있다. 더우기 대부분의 고전적인 방법들에 있어서 주로 이용된 정규분포 가정을 필요로 하지 않는다. 본 연구에서 제안된 방법은 일반화 우도비에 붓스트랩 방법을 적용하여 구성되었다. 모의 실험을 통하여 검정력을 고찰함으로써 제안된 검정방법의 성능을 연구하였다.

  • PDF

Variable Arrangement for Data Visualization

  • Huh, Moon Yul;Song, Kwang Ryeol
    • Communications for Statistical Applications and Methods
    • /
    • 제8권3호
    • /
    • pp.643-650
    • /
    • 2001
  • Some classical plots like scatterplot matrices and parallel coordinates are valuable tools for data visualization. These tools are extensively used in the modern data mining softwares to explore the inherent data structure, and hence to visually classify or cluster the database into appropriate groups. However, the interpretation of these plots are very sensitive to the arrangement of variables. In this work, we introduce two methods to arrange the variables for data visualization. First method is based on the work of Wegman (1999), and this is to arrange the variables using minimum distance among all the pairwise permutation of the variables. Second method is using the idea of principal components. We Investigate the effectiveness of these methods with parallel coordinates using real data sets, and show that each of the two proposed methods has its own strength from different aspects respectively.

  • PDF

웨이브렛을 이용한 공간적 영역분할에 의한 얼굴 인식 (Wavelet-Based Face Recognition by Divided Area)

  • 이성록;이상효;조창호;조도현;이상철
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2003년도 하계종합학술대회 논문집 Ⅳ
    • /
    • pp.2307-2310
    • /
    • 2003
  • In this paper, a method for face recognition based on the wavelet packet decomposition is proposed. In the proposed method, the input image is decomposed by the 2-level wavelet packet transformation and then the face areas are defined by the Integral Projection technique applied to each of the 1-level subband images, HL and LH. After the defined face areas are divided into three areas, called top, bottom, and border, the mean and the variance of the three areas of the approximation image are computed, and the variance of the single predetermined face area for the rest of 15 detail images, from which the feature vectors of statistical measure are extracted. In this paper we use the wavelet packet decomposition, a generalization of the classical wavelet decomposition, to obtain its richer signal analysis features such as discontinuity in higher derivatives, self-similarity, etc. And we have shown that even with very simple statistical features such as mean values and variance we can make an excellent basis for face classification, if an appropriate probability distance is used.

  • PDF

ASSESSING POPULATION BIOEQUIVALENCE IN A $2{\times}2$ CROSSOVER DESIGN WITH CARRYOVER EFFECT IN A BAYESIAN PERSPECTIVE

  • Oh Hyun-Sook
    • Journal of the Korean Statistical Society
    • /
    • 제35권3호
    • /
    • pp.239-250
    • /
    • 2006
  • A $2{\times}2$ crossover design including carryover effect is considered for assessment of population bioequivalence of two drug formulations in a Bayesian framework. In classical analysis, it is complex to deal with the carryover effect since the estimate of the drug effect is biased in the presence of a carryover effect. The proposed method in this article uses uninformative priors and vague proper priors for objectiveness of priors and the posterior probability distribution of the parameters of interest is derived with given priors. The posterior probabilities of the hypotheses for assessing population bioequivalence are evaluated based on a Markov chain Monte Carlo simulation method. An example with real data set is given for illustration.

산-염기 적정 시스템의 비선형 회귀분석에 관한 고찰 (Nonlinear Regression Analysis of Acid-Base Titration System)

  • 박정오;홍재진
    • 대한임상검사과학회지
    • /
    • 제40권1호
    • /
    • pp.18-25
    • /
    • 2008
  • In classical titrimetric analyses, the major concern is the concentration of titrant, usually the aqueous solution of hydrochloric acid or sodium hydroxide, that could be changed as time goes by and it is accompanied with the inaccuracy of the resulting data. And the statistical approach, the nonlinear regression analysis, which is a well-known statistical method, was introduced to determine the accurate concentration of the titrant and the exact value of parameters, $K_a$, r, $C_a$, $C_b$, for 0.01 M aqueous solutions of analytes, sodium pyruvate, sodium acetate, sodium bicarbonate, ammonium hydroxide, ammonium chloride and acetic acid at $25^{\circ}C$. We used Gauss-Newton method for the linearlization of the nonlinear titration system and the two-parameter fitting showed appreciable convergent data for the parameters of the analytes set with the various range of $K_a$ value.

  • PDF