• 제목/요약/키워드: statistical approach

검색결과 2,344건 처리시간 0.028초

Investigations into Coarsening Continuous Variables

  • Jeong, Dong-Myeong;Kim, Jay-J.
    • 응용통계연구
    • /
    • 제23권2호
    • /
    • pp.325-333
    • /
    • 2010
  • Protection against disclosure of survey respondents' identifiable and/or sensitive information is a prerequisite for statistical agencies that release microdata files from their sample surveys. Coarsening is one of popular methods for protecting the confidentiality of the data. Grouped data can be released in the form of microdata or tabular data. Instead of releasing the data in a tabular form only, having microdata available to the public with interval codes with their representative values greatly enhances the utility of the data. It allows the researchers to compute covariance between the variables and build statistical models or to run a variety of statistical tests on the data. It may be conjectured that the variance of the interval data is lower that of the ungrouped data in the sense that the coarsened data do not have the within interval variance. This conjecture will be investigated using the uniform and triangular distributions. Traditionally, midpoint is used to represent all the values in an interval. This approach implicitly assumes that the data is uniformly distributed within each interval. However, this assumption may not hold, especially in the last interval of the economic data. In this paper, we will use three distributional assumptions - uniform, Pareto and lognormal distribution - in the last interval and use either midpoint or median for other intervals for wage and food costs of the Statistics Korea's 2006 Household Income and Expenditure Survey(HIES) data and compare these approaches in terms of the first two moments.

통계적 편차와 히스토그램 변형을 이용한 단일영상기반 고품질 영상 생성기법 (Single Image Based HDR Algorithm Using Statistical Differencing and Histogram Manipulation)

  • 송진선;한규필;박양우
    • 한국멀티미디어학회논문지
    • /
    • 제21권7호
    • /
    • pp.764-771
    • /
    • 2018
  • In this paper, we propose a high-quality image acquisition algorithm using only a single image, which the high-quality image is normally referred as HDR ones. In order to acquire the HDR image, conventional methods need many images having different exposure values at the same scene and should delicately adjust the color values for a bit-expansion or an exposure fusion. Thus, they require considerable calculations and complex structures. Therefore, the proposed algorithm suggests a completely new approach using one image for the high-quality image acquisition by applying statistical difference and histogram manipulation, or histogram specification, techniques. The techniques could control the pixel's statistical distribution of the input image into the desired one through the local and the global modifications, respectively. As the result, the quality of the proposed algorithm is better than those of conventional methods implemented in commercial image editing softwares.

와이블 고장모형 하에서의 이중샘플링 T2 관리도의 경제적-통계적 설계 (이중샘플링 T2 관리도의 경제적-통계적 설계) (Economic-Statistical Design of Double Sampling T2 Control Chart under Weibull Failure Model)

  • 홍성옥;이민구;이주호
    • 품질경영학회지
    • /
    • 제43권4호
    • /
    • pp.471-488
    • /
    • 2015
  • Purpose: Double sampling $T^2$ chart is a useful tool for detecting a relatively small shift in process mean when the process is controlled by multiple variables. This paper finds the optimal design of the double sampling $T^2$ chart in both economical and statistical sense under Weibull failure model. Methods: The expected cost function is mathematically derived using recursive equation approach. The optimal designs are found using a genetic algorithm for numerical examples and compared to those of single sampling $T^2$ chart. Sensitivity analysis is performed to see the parameter effects. Results: The proposed design outperforms the optimal design of the single sampling $T^2$ chart in terms of the expected cost per unit time and Type-I error rate for all the numerical examples considered. Conclusion: Double sampling $T^2$ chart can be designed to satisfy both economic and statistical requirements under Weibull failure model and the resulting design is better than the single sampling counterpart.

STATISTICAL PROPERTIES OF GRAVITATIONAL LENSING IN COSMOLOGICAL MODELS WITH COSMOLOGICAL CONSTANT

  • LEE HYUN-A;PARK MYEONG-GU
    • 천문학회지
    • /
    • 제27권2호
    • /
    • pp.103-117
    • /
    • 1994
  • To extend the work of Gott, Park, and Lee (1989), statistical properties of gravitational lensing in a wide variety of cosmological models involving non-zero cosmological constant is investigated, using the redshifts of both lens and source and observed angular separation of images for gravitational lens systems. We assume singular isothermal sphere as lensing galaxy in homogenous and isotropic Friedmann­Lemaitre-Robertson- Walker universe, Schechter luminosity function, standard angular diameter distance formula and other galaxy parameters used in Fukugita and Turner (1991). To find the most adequate flat cosmological model and put a limit on the value of dimensionless cosmological constant $\lambda_0$, the mean value of the angular separation of images, probability distribution of angular separation and cumulative probability are calculated for given source and lens redshifts and compared with the observed values through several statistical methods. When there is no angular selection effect, models with highest value of $\lambda_0$ is preferred generally. When the angular selection effects are considered, the preferred model depends on the shape of the selection functions and statistical methods; yet, models with large $\lambda_0$ are preferred in general. However, the present data can not rule out any of the flat universe models with enough confidence. This approach can potentially select out best model. But at the moment, we need more data.

  • PDF

온라인 서포트벡터기계를 이용한 온라인 비정상 사건 탐지 (Online abnormal events detection with online support vector machine)

  • 박혜정
    • Journal of the Korean Data and Information Science Society
    • /
    • 제22권2호
    • /
    • pp.197-206
    • /
    • 2011
  • 신호처리 관련 응용문제에서는 신호에서 실시간으로 발생하는 비정상적인 사건들을 탐지하는 것이 매우 중요하다. 이전에 알려져 있는 비정상 사건 탐지방법들은 신호에 대한 명확한 통계적인 모형을 가정하고, 비정상적인 신호들은 통계적인 모형의 가정 하에서 비정상적인 사건들로 해석한다. 탐지방법으로 최대우도와 베이즈 추정 이론이 많이 사용되고 있다. 그러나 앞에서 언급한 방법으로는 로버스트 하고 다루기 쉬운 모형을 추정한다는 것은 쉽지가 않다. 좀 더 로버스트한 모형을 추정할 수 있는 방법이 필요하다. 본 논문에서는 로버스트 하다고 알려져 있는 서포트 벡터 기계를 이용하여 온라인으로 비정상적인 신호를 탐지하는 방법을 제안한다.

HMM 기반의 TTS를 위한 상호유사도 비율을 이용한 결정트리 기반의 문맥 군집화 (Decision Tree Based Context Clustering with Cross Likelihood Ratio for HMM-based TTS)

  • 정치상;강홍구
    • 한국음향학회지
    • /
    • 제32권2호
    • /
    • pp.174-180
    • /
    • 2013
  • 본 논문은 HMM 기반의 TTS 시스템을 위하여 상호유사도 비율을 이용한 결정트리 기반의 문맥 군집화 알고리즘을 제안한다. 기존의 알고리즘들은 유사한 통계적 특성을 가지는 문맥종속 HMM을 하나로 묶고 있다. 그러나 기존의 알고리즘들은 결정트리의 나누어진 노드간의 통계적 유사도를 고려하지 않음으로 인하여 최종 노드 사이의 통계적인 차이를 보장하지 못한다. 제안한 알고리즘은 분리된 노드들 간의 통계적 유사도를 최소화하여 모델 파라미터의 신뢰도를 향상시킨다. 실험 결과를 통해 제안한 알고리즘이 기존의 알고리즘들에 비해 우수한 성능을 나타낸다는 것을 확인할 수 있다.

Application of Bayesian Statistical Analysis to Multisource Data Integration

  • Hong, Sa-Hyun;Moon, Wooil-M.
    • 대한원격탐사학회:학술대회논문집
    • /
    • 대한원격탐사학회 2002년도 Proceedings of International Symposium on Remote Sensing
    • /
    • pp.394-399
    • /
    • 2002
  • In this paper, Multisource data classification methods based on Bayesian formula are considered. For this decision fusion scheme, the individual data sources are handled separately by statistical classification algorithms and then Bayesian fusion method is applied to integrate from the available data sources. This method includes the combination of each expert decisions where the weights of the individual experts represent the reliability of the sources. The reliability measure used in the statistical approach is common to all pixels in previous work. In this experiment, the weight factors have been assigned to have different value for all pixels in order to improve the integrated classification accuracies. Although most implementations of Bayesian classification approaches assume fixed a priori probabilities, we have used adaptive a priori probabilities by iteratively calculating the local a priori probabilities so as to maximize the posteriori probabilities. The effectiveness of the proposed method is at first demonstrated on simulations with artificial and evaluated in terms of real-world data sets. As a result, we have shown that Bayesian statistical fusion scheme performs well on multispectral data classification.

  • PDF

Statistical damage classification method based on wavelet packet analysis

  • Law, S.S.;Zhu, X.Q.;Tian, Y.J.;Li, X.Y.;Wu, S.Q.
    • Structural Engineering and Mechanics
    • /
    • 제46권4호
    • /
    • pp.459-486
    • /
    • 2013
  • A novel damage classification method based on wavelet packet transform and statistical analysis is developed in this study for structural health monitoring. The response signal of a structure under an impact load is normalized and then decomposed into wavelet packet components. Energies of these wavelet packet components are then calculated to obtain the energy distribution. Statistical similarity comparison based on an F-test is used to classify the structure from changes in the wavelet packet energy distribution. A statistical indicator is developed to describe the damage extent of the structure. This approach is applied to the test results from simply supported reinforced concrete beams in the laboratory. Cases with single and two damages are created from static loading, and accelerations of the structure from under impact loads are analyzed. Results show that the method can be used with no reference baseline measurement and model for the damage monitoring and assessment of the structure with alarms at a specified significance level.

오스테나이트계 스테인리스강 노내 구조물의 조사유기응력부식균열 영향 인자에 대한 통계적 분석 (Statistical Evaluation of Factors Affecting IASCC of Austenitic Stainless Steels for PWR Core Internals)

  • 김성우;황성식;김홍표
    • 대한금속재료학회지
    • /
    • 제47권12호
    • /
    • pp.819-827
    • /
    • 2009
  • This work is concerned with a statistical analysis of factors affecting the irradiation-assisted stress corrosion cracking (IASCC) of austenitic stainless steels for core internals of pressurized water reactors (PWR). The microstructural and environmental factors were reviewed and critically evaluated by the statistical analysis. The Cr depletion at grain boundary was determined to have no significant correlation with the IASCC susceptibility. The threshold irradiation fluence of IASCC in a PWR was statistically calculated to decrease from 5.799 to 1.914 DPA with increase of temperature from 320 to $340^{\circ}C$. From the analysis of the relationship between applied stress and time-to-failure of stainless steel components based on an accelerated life testing model, it was found that B2 life of a baffle former bolt exposed to neutron fluence of 20 and 75 DPA was at least 2.5 and 0.4 year, respectively, within 95% confidence interval.

A pooled Bayes test of independence using restricted pooling model for contingency tables from small areas

  • Jo, Aejeong;Kim, Dal Ho
    • Communications for Statistical Applications and Methods
    • /
    • 제29권5호
    • /
    • pp.547-559
    • /
    • 2022
  • For a chi-squared test, which is a statistical method used to test the independence of a contingency table of two factors, the expected frequency of each cell must be greater than 5. The percentage of cells with an expected frequency below 5 must be less than 20% of all cells. However, there are many cases in which the regional expected frequency is below 5 in general small area studies. Even in large-scale surveys, it is difficult to forecast the expected frequency to be greater than 5 when there is small area estimation with subgroup analysis. Another statistical method to test independence is to use the Bayes factor, but since there is a high ratio of data dependency due to the nature of the Bayesian approach, the low expected frequency tends to decrease the precision of the test results. To overcome these limitations, we will borrow information from areas with similar characteristics and pool the data statistically to propose a pooled Bayes test of independence in target areas. Jo et al. (2021) suggested hierarchical Bayesian pooling models for small area estimation of categorical data, and we will introduce the pooled Bayes factors calculated by expanding their restricted pooling model. We applied the pooled Bayes factors using bone mineral density and body mass index data from the Third National Health and Nutrition Examination Survey conducted in the United States and compared them with chi-squared tests often used in tests of independence.