• Title/Summary/Keyword: Sample Size

Search Result 3,903, Processing Time 0.029 seconds

The Influence of Sample Size on Environment Assessment Using Benthic Macroinvertebrates (저서성 대형무척추동물을 이용한 환경평가에서 표본크기가 미치는 영향)

  • Kim, Ah Reum;Oh, Min Woo;Kong, Dongsoo
    • Journal of Korean Society on Water Environment
    • /
    • v.29 no.6
    • /
    • pp.790-798
    • /
    • 2013
  • Benthic macroinvertebrates are widely used as biological indicators for assessing the integrity of aquatic ecosystem. However, sampling has usually been done with fixed sample size due to time consuming and costly process. This study was conducted to find out the influence of sample size on the biological indices (H' DI, R1, J, EPT, ESB and BMI) of benthic macroinvertebrates. The 15 replicate samples were quantitatively collected from each 3 different site of two mountain streams in May, 2011. With the replicate data, we combined the abundance of each species with all the possible combinations of the sample size. Along with the increase of sample size, the number of species increased continuously and did not converge. BMI showed little difference whereas other biological indices increased or decreased.

Sample size calculations for clustered count data based on zero-inflated discrete Weibull regression models

  • Hanna Yoo
    • Communications for Statistical Applications and Methods
    • /
    • v.31 no.1
    • /
    • pp.55-64
    • /
    • 2024
  • In this study, we consider the sample size determination problem for clustered count data with many zeros. In general, zero-inflated Poisson and binomial models are commonly used for zero-inflated data; however, in real data the assumptions that should be satisfied when using each model might be violated. We calculate the required sample size based on a discrete Weibull regression model that can handle both underdispersed and overdispersed data types. We use the Monte Carlo simulation to compute the required sample size. With our proposed method, a unified model with a low failure risk can be used to cope with the dispersed data type and handle data with many zeros, which appear in groups or clusters sharing a common variation source. A simulation study shows that our proposed method provides accurate results, revealing that the sample size is affected by the distribution skewness, covariance structure of covariates, and amount of zeros. We apply our method to the pancreas disorder length of the stay data collected from Western Australia.

An Overview of Bootstrapping Method Applicable to Survey Researches in Rehabilitation Science

  • Choi, Bong-sam
    • Physical Therapy Korea
    • /
    • v.23 no.2
    • /
    • pp.93-99
    • /
    • 2016
  • Background: Parametric statistical procedures are typically conducted under the condition in which a sample distribution is statistically identical with its population. In reality, investigators use inferential statistics to estimate parameters based on the sample drawn because population distributions are unknown. The uncertainty of limited data from the sample such as lack of sample size may be a challenge in most rehabilitation studies. Objects: The purpose of this study is to review the bootstrapping method to overcome shortcomings of limited sample size in rehabilitation studies. Methods: Articles were reviewed. Results: Bootstrapping method is a statistical procedure that permits the iterative re-sampling with replacement from a sample when the population distribution is unknown. This statistical procedure is to enhance the representativeness of the population being studied and to determine estimates of the parameters when sample size are too limited to generalize the study outcome to target population. The bootstrapping method would overcome limitations such as type II error resulting from small sample sizes. An application on a typical data of a study represented how to deal with challenges of estimating a parameter from small sample size and enhance the uncertainty with optimal confidence intervals and levels. Conclusion: Bootstrapping method may be an effective statistical procedure reducing the standard error of population parameters under the condition requiring both acceptable confidence intervals and confidence level (i.e., p=.05).

Efficient determination of the size of experiments by using graphs in balanced design of experiments (균형된 실험계획법에서 그래프를 활용한 실험의 크기의 효율적인 결정)

  • Lim, Yong B.;Youn, Sora;Chung, Jong Hee
    • Journal of Korean Society for Quality Management
    • /
    • v.46 no.3
    • /
    • pp.651-658
    • /
    • 2018
  • Purpose: The algorithm described in Lim(1998) is available to determine the sample size directly given specified significance level, power and signal-to-noise ratio. We research on the efficient determination of the sample size by visual methods. Methods: We propose three graphs for investigating the mutual relationship between the sample size r, power $1-{\beta}$ and the detectable signal-to-noise ratio ${\Delta}$. First graph shows the relationship between ${\Delta}$ and $1-{\beta}$ for the given r and it can be checked whether the power is sufficient enough. Second graph shows the relationship between r and ${\Delta}$ for the given power $1-{\beta}$. Third graph shows the relationship between r and $1-{\beta}$ for the given ${\Delta}$. It can be checked that which effects are sensitive to the efficient sample size by investigating those graphs. Results: In factorial design, randomized block design and the split plot design how to determine the sample size directly given specified significance level, power and signal-to-noise ratio is programmed by using R. A experiment to study the split plot design in Hicks(1982) is used as an example. We compare the sample sizes calculated by randomized block design with those by split plot design. By using graphs, we can check the possibility of reducing the sample size efficiently. Conclusion: The proposed visual methods can help an engineer to make a proper plan to reduce the sample size.

Sample size calculation for comparing time-averaged responses in K-group repeated binary outcomes

  • Wang, Jijia;Zhang, Song;Ahn, Chul
    • Communications for Statistical Applications and Methods
    • /
    • v.25 no.3
    • /
    • pp.321-328
    • /
    • 2018
  • In clinical trials with repeated measurements, the time-averaged difference (TAD) may provide a more powerful evaluation of treatment efficacy than the rate of changes over time when the treatment effect has rapid onset and repeated measurements continue across an extended period after a maximum effect is achieved (Overall and Doyle, Controlled Clinical Trials, 15, 100-123, 1994). The sample size formula has been investigated by many researchers for the evaluation of TAD in two treatment groups. For the evaluation of TAD in multi-arm trials, Zhang and Ahn (Computational Statistics & Data Analysis, 58, 283-291, 2013) and Lou et al. (Communications in Statistics-Theory and Methods, 46, 11204-11213, 2017b) developed the sample size formulas for continuous outcomes and count outcomes, respectively. In this paper, we derive a sample size formula to evaluate the TAD of the repeated binary outcomes in multi-arm trials using the generalized estimating equation approach. This proposed sample size formula accounts for various correlation structures and missing patterns (including a mixture of independent missing and monotone missing patterns) that are frequently encountered by practitioners in clinical trials. We conduct simulation studies to assess the performance of the proposed sample size formula under a wide range of design parameters. The results show that the empirical powers and the empirical Type I errors are close to nominal levels. We illustrate our proposed method using a clinical trial example.

Effects of the Reference Sample Size on the Performance of the Two-Sample Rank Detector (두 표본 순위 검파에서 기준 표본 크기가 검파기 성능에 미치는 영향)

  • Bae, Jinsoo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.8
    • /
    • pp.1515-1517
    • /
    • 2015
  • The effects of the reference sample size on the detection probability of the two-sample rank detector is investigated in this paper. The larger reference sample size shows the better performance of the detector. The effect is also shown to be saturated as the reference sample size becomes larger.

A sample size calibration approach for the p-value problem in huge samples

  • Park, Yousung;Jeon, Saebom;Kwon, Tae Yeon
    • Communications for Statistical Applications and Methods
    • /
    • v.25 no.5
    • /
    • pp.545-557
    • /
    • 2018
  • The inclusion of covariates in the model often affects not only the estimates of meaningful variables of interest but also its statistical significance. Such gap between statistical and subject-matter significance is a critical issue in huge sample studies. A popular huge sample study, the sample cohort data from Korean National Health Insurance Service, showed such gap of significance in the inference for the effect of obesity on cause of mortality, requiring careful consideration. In this regard, this paper proposes a sample size calibration method based on a Monte Carlo t (or z)-test approach without Monte Carlo simulation, and also proposes a test procedure for subject-matter significance using this calibration method in order to complement the deflated p-value in the huge sample size. Our calibration method shows no subject-matter significance of the obesity paradox regardless of race, sex, and age groups, unlike traditional statistical suggestions based on p-values.

Sample Size Calculations with Dropouts in Clinical Trials (임상시험에서 중도탈락을 고려한 표본크기의 결정)

  • Lee, Ki-Hoon
    • Communications for Statistical Applications and Methods
    • /
    • v.15 no.3
    • /
    • pp.353-365
    • /
    • 2008
  • The sample size in a clinical trial is determined by the hypothesis, the variance of observations, the effect size, the power and the significance level. Dropouts in clinical trials are inevitable, so we need to consider dropouts on the determination of sample size. It is common that some proportion corresponding to the expected dropout rate would be added to the sample size calculated from a mathematical equation. This paper proposes new equations for calculating sample size dealing with dropouts. Since we observe data longitudinally in most clinical trials, we can use a last observation to impute for missing one in the intention to treat (ITT) trials, and this technique is called last observation carried forward(LOCF). But LOCF might make deviations on the assumed variance and effect size, so that we could not guarantee the power of test with the sample size obtained from the existing equation. This study suggests the formulas for sample size involving information about dropouts and shows the properties of the proposed method in testing equality of means.

Group Control Charts with Variable Stream and Sample Sizes (가변 스트림 및 표본크기 그룹관리도)

  • Lee, K.T.;Bai, D.S.
    • Journal of Korean Institute of Industrial Engineers
    • /
    • v.24 no.3
    • /
    • pp.333-343
    • /
    • 1998
  • This paper proposes variable stream and sample size(VSSS) group control charts in which both the number of streams selected for sampling and sample size from each of the selected streams are allowed to vary based on the values of the preceding sample statistics. The proposed charts select a small portion of streams and take samples of size n = 1 if both the largest and smallest of sample means fall between the lower and upper threshold limits, and select a large portion of streams and take samples of size n > 1 otherwise. A Markov chain approach is used to derive the formulas for evaluating the performances of the proposed charts. Numerical comparisons are made between the VSSS and fixed stream and sample size(FSSS) group control charts.

  • PDF

Sample Size Determination for O/D Estimation under Budget Constraint (예산제약하에서 O/D 추정을 위한 최소표본율 결정)

  • Sin, Hui-Cheol;Lee, Hyang-Suk
    • Journal of Korean Society of Transportation
    • /
    • v.24 no.3 s.89
    • /
    • pp.7-15
    • /
    • 2006
  • A large sample can Provide more information about the Population. As the sample size Increases, analysts will be more confident about the survey results. On the other hand, the costs for survey will increase in time and manpower. Therefore, determination of the sample size is a trade-off between the required accuracy and the cost. In addition, permitted error and significance level should be considered. Sample size determination in surveys for O/D estimation is also connected with confidence of survey result. However, the past methods were usually too simple to consider confidence. Therefore, a new method for O/D surveys was Proposed and it was accurate enough, but it has too large sample size when we have current budget constraint. In this research, several minimum sample size determination methods for origin-destination survey under budget constraint were proposed. Each method decreased sample size, but has its own advantages. Selection of the sample size will depend on the study Purpose and budget constraint.