• 제목/요약/키워드: data-fitting

검색결과 1,451건 처리시간 0.029초

Analysis of Quasi-Likelihood Models using SAS/IML

  • Ha, Il-Do
    • Journal of the Korean Data and Information Science Society
    • /
    • 제8권2호
    • /
    • pp.247-260
    • /
    • 1997
  • The quasi-likelihood models which greatly widened the scope of generalized linear models are widely used in data analysis where a likelihood is not available. Since a quasi-likelihood may not appear to be an ordinary likelihood for any known distribution in the natural exponential family, to fit the quasi-likelihood models the standard statistical packages such as GLIM, GENSTAT, S-PLUS and so on may not directly applied. SAS/IML is very useful for fitting of such models. In this paper, we present simple SAS/IML(version 6.11) program which helps to fit and analyze the quasi-likelihood models applied to the leaf-blotch data introduced by Wedderburn(1974), and the problem with deviance useful generally to model checking is pointed out, and then its solution method is mention through the data analysis based on this quasi-likelihood models checking.

  • PDF

A Study of HME Model in Time-Course Microarray Data

  • Myoung, Sung-Min;Kim, Dong-Geon;Jo, Jin-Nam
    • 응용통계연구
    • /
    • 제25권3호
    • /
    • pp.415-422
    • /
    • 2012
  • For statistical microarray data analysis, clustering analysis is a useful exploratory technique and offers the promise of simultaneously studying the variation of many genes. However, most of the proposed clustering methods are not rigorously solved for a time-course microarray data cluster and for a fitting time covariate; therefore, a statistical method is needed to form a cluster and represent a linear trend of each cluster for each gene. In this research, we developed a modified hierarchical mixture of an experts model to suggest clustering data and characterize each cluster using a linear mixed effect model. The feasibility of the proposed method is illustrated by an application to the human fibroblast data suggested by Iyer et al. (1999).

Analyzing Survival Data as Binary Outcomes with Logistic Regression

  • Lim, Jo-Han;Lee, Kyeong-Eun;Hahn, Kyu-S.;Park, Kun-Woo
    • Communications for Statistical Applications and Methods
    • /
    • 제17권1호
    • /
    • pp.117-126
    • /
    • 2010
  • Clinical researchers often analyze survival data as binary outcomes using the logistic regression method. This paper examines the information loss resulting from analyzing survival time as binary outcomes. We first demonstrate that, under the proportional hazard assumption, this binary discretization does result in a significant information loss. Second, when fitting a logistic model to survival time data, researchers inadvertently use the maximal statistic. We implement a numerical study to examine the properties of the reference distribution for this statistic, finally, we show that the logistic regression method can still be a useful tool for analyzing survival data in particular when the proportional hazard assumption is questionable.

Testing the exchange rate data for the parameter change based on ARMA-GARCH model

  • Song, Junmo;Ko, Bangwon
    • Journal of the Korean Data and Information Science Society
    • /
    • 제24권6호
    • /
    • pp.1551-1559
    • /
    • 2013
  • In this paper, we analyze the Korean Won/Japanese 100 Yen exchange rate data based on the ARMA-GARCH model, and perform the test for detecting the parameter changes. As a test statistics, we employ the cumulative sum (CUSUM) test for ARMA-GARCH model, which is introduced by Lee and Song (2008). Our empirical analysis indicates that the KRW/JPY exchange rate series experienced several parameter changes during the period from January 2000 to December 2012, which leads to a fitting of AR-IGARCH model to the whole series.

Probabilistic Approach on Railway Infrastructure Stability and Settlement Analysis

  • Lee, Sangho
    • International Journal of Railway
    • /
    • 제6권2호
    • /
    • pp.45-52
    • /
    • 2013
  • Railway construction needs vast soil investigation for its infrastructure foundation designs along the planned railway path to identify the design parameters for stability and serviceability checks. The soil investigation data are usually classified and grouped to decide design input parameters per each construction section and budget estimates. Deterministic design method which most civil engineer and practitioner are familiar with has a clear limitation in construction/maintenance budget control, and occasionally produced overdesigned or unsafe design problems. Instead of using a batch type analysis with predetermined input parameters, data population collected from site soil investigation and design load condition can be statistically estimated for the mean and variance to present the feature of data distribution and optimized with a best fitting probability function. Probabilistic approach using entire feature of design input data enables to predict the worst, best and most probable cases based on identified ranges of soil and load data, which will help railway designer select construction method to save the time and cost. This paper introduces two Monte Carlo simulations actually applied on estimation of retaining wall external stability and long term settlement of organic soil in soil investigation area for a recent high speed railway project.

Cubic normal distribution and its significance in structural reliability

  • Zhao, Yan-Gang;Lu, Zhao-Hui
    • Structural Engineering and Mechanics
    • /
    • 제28권3호
    • /
    • pp.263-280
    • /
    • 2008
  • Information on the distribution of the basic random variable is essential for the accurate analysis of structural reliability. The usual method for determining the distributions is to fit a candidate distribution to the histogram of available statistical data of the variable and perform approximate goodness-of-fit tests. Generally, such candidate distribution would have parameters that may be evaluated from the statistical moments of the statistical data. In the present paper, a cubic normal distribution, whose parameters are determined using the first four moments of available sample data, is investigated. A parameter table based on the first four moments, which simplifies parameter estimation, is given. The simplicity, generality, flexibility and advantages of this distribution in statistical data analysis and its significance in structural reliability evaluation are discussed. Numerical examples are presented to demonstrate these advantages.

Type 316LN 강의 크리프 수명예측 파라메타의 표준오차 분석 (Standard Error Analysis of Creep-Life Prediction Parameters of Type 316LN Stainless Steels)

  • 김우곤;윤송남;류우석
    • 대한기계학회:학술대회논문집
    • /
    • 대한기계학회 2004년도 춘계학술대회
    • /
    • pp.19-24
    • /
    • 2004
  • A number of creep data were collected and filed for type 316LN stainless steels through literature survey and experimental data produced in KAERI. Using these data, polynomial equations for predicting creep life were obtained for Larson Miller (L-M), Qrr-Sherby-Dorn (O-S-D) and Manson-Haferd (M-H) parametric methods. In order to find out the suitability for them, the relative standard error (RSE) and standard error of estimate (SEE) values were obtained by statistical process of creep data. The O-S-D parameter showed better fitting to creep-rupture data than the L-M or the M-H parameters, and the three parametric methods did not generate the large difference in the SEE and the RSE values.

  • PDF

분산 자료에 대한 초완비 표현 방법 (A method of overcomplete representation for distributed data)

  • 이상철;박종우;곽칠성
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2007년도 하계종합학술대회 논문집
    • /
    • pp.457-458
    • /
    • 2007
  • This paper propose a method for representing distributed data of sensor networks. The proposed method is based on a general distributed regression framework that models sensor data by fitting a global function to each of the local measurements and explores the possible extensions of distribution regression by using complex signal representations. In order to reduce the amount of processed data and the required communication, the signal model has to be as compact as possible, with the ability to restore the arbitrary measurements. To achieve this requirement, data compression step is included, where the basis function set is changed to an overcomplete set. This have better advantages in case of nonstationary signal modeling than complete base representation.

  • PDF

Philips LINAC 6 MV와 8 MV X선 소조사연에 대한 선량분포 측정 (Measurement of Dose Distribution in Small Beams of Philips 6 and 8 MVX Linear Accelerator)

  • 서태석;윤세철;신경섭;박용휘
    • Radiation Oncology Journal
    • /
    • 제9권1호
    • /
    • pp.143-152
    • /
    • 1991
  • 본 논문에서는 소조사면에 대한 X-선의 선량분포를 일반실험식으로 계산될 수 있도록 beam 측정 데이타를 종합 처리하는 방법에 대하여 기술하고 있다. Beam 데이타는 philips LINAC 6 MV, 8 MV X-ray에 대하여 측정 되었으며, 측정된 요소는 tissue maximum ratio (TMR), off-axis-ratio (OAH), 그리고 relative output factor (ROF)를 포함한다. 소조사면에 의한 방사선 치료를 위하여 isocenter에서 지름이 1 내지 3cm되도록 실린더 형태의 특수 collimator가 2 mm 간격으로 제작되었다. 본 측정을 위하여 다이오드 detector가 이용되었으며 Film 및 TLO 측정기로 측정된 값과 비교검토 되었다. 제한된 조사면으로 측정된 TMR, OAR data로부터 beam 데이타를 나타내는 실험식을 유도하였으며 이 실험식은 임의의 Set-UP조건에 따른 측정값을 예상할 수 있는 일반 실험식으로 확장되었고 측정된 TMR과 OAR 값들은 잘 일치되었다.

  • PDF

발파 진동식의 신뢰성 (The Reliability of Blast Vibration Equation)

  • 김수일;정상섬;조후연
    • 대한토목학회논문집
    • /
    • 제14권3호
    • /
    • pp.573-582
    • /
    • 1994
  • 본 논문에서는 이미 제안된 발파진동식 중에서 국내의 지질조건에 가장 적합한 식을 연구하였다. 국내에서 측정된 여러 현장의 자료를 이용하여 제안된 발파진동식의 적합성을 분석 검토하였다. 실측자료를 이용한 발파진동식의 산정은 선형회귀분석을 적용하였다. 또한 실측자료로 각 발파진동식을 산출한 후에는 이 발파진동식에 다시 환산거리를 대입하여 진동속도를 산출하였다. 산출한 진동속도와 측정한 진동속도를 비교함으로써 회귀분석한 발파진동식의 신뢰성을 도심지의 소규모발파와 채석장의 대규모발파를 나누어서 살펴보았다. 그 결과 국내의 지질조건에 가장 적합한 식은 미광무국에서 제안한 ROOT SCALE과 CUBE ROOT SCALE 임을 밝혔다. 또한 본 논문에서는 실측자료와 기존의 현장자료를 이용하여 각 암종을 대표할 수 있는 발파진동식을 제안하였다.

  • PDF