• Title/Summary/Keyword: Statistical Analysis Data

Search Result 9,183, Processing Time 0.037 seconds

A Study on the Validity of the Statistical Collection and Analysis in Gwangju and Chonnam (통계자료의 수집 및 분석의 타당성에 관한 연구- 광주,전남지역을 중심으로 -)

  • 이화영
    • The Korean Journal of Applied Statistics
    • /
    • v.6 no.2
    • /
    • pp.443-452
    • /
    • 1993
  • A check list which includes the items that are to be considered in the process of the statistical data collection and analysis by non-scientific organizations is proposed. Based on the suggested check list, the output resulting from the statistical survey conducted by private organizations, banks, organs of expression and enterprises in Gwangju and Chonnam are examined about the validity of data collection and statistical analysis.

  • PDF

Radioactive waste sampling for characterisation - A Bayesian upgrade

  • Pyke, Caroline K.;Hiller, Peter J.;Koma, Yoshikazu;Ohki, Keiichi
    • Nuclear Engineering and Technology
    • /
    • v.54 no.1
    • /
    • pp.414-422
    • /
    • 2022
  • Presented in this paper is a methodology for combining a Bayesian statistical approach with Data Quality Objectives (a structured decision-making method) to provide increased levels of confidence in analytical data when approaching a waste boundary. Development of sampling and analysis plans for the characterisation of radioactive waste often use a simple, one pass statistical approach as underpinning for the sampling schedule. Using a Bayesian statistical approach introduces the concept of Prior information giving an adaptive sample strategy based on previous knowledge. This aligns more closely with the iterative approach demanded of the most commonly used structured decision-making tool in this area (Data Quality Objectives) and the potential to provide a more fully underpinned justification than the more traditional statistical approach. The approach described has been developed in a UK regulatory context but is translated to a waste stream from the Fukushima Daiichi Nuclear Power Station to demonstrate how the methodology can be applied in this context to support decision making regarding the ultimate disposal option for radioactive waste in a more global context.

Unequal Size, Two-way Analysis of Variance for Categorical Data

  • Chung, Han-Yong
    • Journal of the Korean Statistical Society
    • /
    • v.5 no.1
    • /
    • pp.29-34
    • /
    • 1976
  • The techniques about the analysis of variance for quantitative variables have been well-developed. But when the variable is categorical, we must switch to a completely different set of varied techniques. R.J. Light and B.H. Margolin presented one kind of techniques for categorical data in their paper, where there are G unordered experimental groups and I unordered response categories.

  • PDF

Results of Discriminant Analysis with Respect to Cluster Analyses Under Dimensional Reduction

  • Chae, Seong-San
    • Communications for Statistical Applications and Methods
    • /
    • v.9 no.2
    • /
    • pp.543-553
    • /
    • 2002
  • Principal component analysis is applied to reduce p-dimensions into q-dimensions ( $q {\leq} p$). Any partition of a collection of data points with p and q variables generated by the application of six hierarchical clustering methods is re-classified by discriminant analysis. From the application of discriminant analysis through each hierarchical clustering method, correct classification ratios are obtained. The results illustrate which method is more reasonable in exploratory data analysis.

Review on statistical methods for protecting privacy and measuring risk of disclosure when releasing information for public use (정보공개 환경에서 개인정보 보호와 노출 위험의 측정에 대한 통계적 방법)

  • Lee, Yonghee
    • Journal of the Korean Data and Information Science Society
    • /
    • v.24 no.5
    • /
    • pp.1029-1041
    • /
    • 2013
  • Recently, along with emergence of big data, there are incresing demands for releasing information and micro data for public use so that protecting privacy and measuring risk of disclosure for released database become important issues in goverment and business sector as well as academic community. This paper reviews statistical methods for protecting privacy and measuring risk of disclosure when micro data or data analysis sever is released for public use.

Bayesian Analysis in Generalized Log-Gamma Censored Regression Model

  • Younshik chung;Yoomi Kang
    • Communications for Statistical Applications and Methods
    • /
    • v.5 no.3
    • /
    • pp.733-742
    • /
    • 1998
  • For industrial and medical lifetime data, the generalized log-gamma regression model is considered. Then the Bayesian analysis for the generalized log-gamma regression with censored data are explained and following the data augmentation (Tanner and Wang; 1987), the censored data is replaced by simulated data. To overcome the complicated Bayesian computation, Makov Chain Monte Carlo (MCMC) method is employed. Then some modified algorithms are proposed to implement MCMC. Finally, one example is presented.

  • PDF

Statistical Mistakes Commonly Made When Writing Medical Articles (의학 논문 작성 시 발생하는 흔한 통계적 오류)

  • Soyoung Jeon;Juyeon Yang;Hye Sun Lee
    • Journal of the Korean Society of Radiology
    • /
    • v.84 no.4
    • /
    • pp.866-878
    • /
    • 2023
  • Statistical analysis is an essential component of the medical writing process for research-related articles. Although the importance of statistical testing is emphasized, statistical mistakes continue to appear in journal articles. Major statistical mistakes can occur in any of the three different stages of medical writing, including in the design stage, analysis stage, and interpretation stage. In the design stage, mistakes occur if there is a lack of specificity regarding the research hypothesis or data collection and analysis plans. Discrepancies in the analysis stage occur if the purpose of the study and characteristics of the data are not sufficiently considered, or when an inappropriate analytic procedure is followed. After performing the analysis, the results are interpreted, and an article is written. Statistical analysis mistakes can occur if the underlying methods are incorrectly written or if the results are misinterpreted. In this paper, we describe the statistical mistakes that commonly occur in medical research-related articles and provide advice with the aim to help readers reduce, resolve, and avoid these mistakes in the future.

A Data Mining Approach for a Dynamic Development of an Ontology-Based Statistical Information System

  • Mohamed Hachem Kermani;Zizette Boufaida;Amel Lina Bensabbane;Besma Bourezg
    • Journal of Information Science Theory and Practice
    • /
    • v.11 no.2
    • /
    • pp.67-81
    • /
    • 2023
  • This paper presents a dynamic development of an ontology-based statistical information system supporting the collection, storage, processing, analysis, and the presentation of statistical knowledge at the national scale. To accomplish this, we propose a data mining technique to dynamically collect data relating to citizens from publicly available data sources; the collected data will then be structured, classified, categorized, and integrated into an ontology. Moreover, an intelligent platform is proposed in order to generate quantitative and qualitative statistical information based on the knowledge stored in the ontology. The main aims of our proposed system are to digitize administrative tasks and to provide reliable statistical information to governmental, economic, and social actors. The authorities will use the ontology-based statistical information system for strategic decision-making as it easily collects, produces, analyzes, and provides both quantitative and qualitative knowledge that will help to improve the administration and management of national political, social, and economic life.

Quantitative Analysis for Plasma Etch Modeling Using Optical Emission Spectroscopy: Prediction of Plasma Etch Responses

  • Jeong, Young-Seon;Hwang, Sangheum;Ko, Young-Don
    • Industrial Engineering and Management Systems
    • /
    • v.14 no.4
    • /
    • pp.392-400
    • /
    • 2015
  • Monitoring of plasma etch processes for fault detection is one of the hallmark procedures in semiconductor manufacturing. Optical emission spectroscopy (OES) has been considered as a gold standard for modeling plasma etching processes for on-line diagnosis and monitoring. However, statistical quantitative methods for processing the OES data are still lacking. There is an urgent need for a statistical quantitative method to deal with high-dimensional OES data for improving the quality of etched wafers. Therefore, we propose a robust relevance vector machine (RRVM) for regression with statistical quantitative features for modeling etch rate and uniformity in plasma etch processes by using OES data. For effectively dealing with the OES data complexity, we identify seven statistical features for extraction from raw OES data by reducing the data dimensionality. The experimental results demonstrate that the proposed approach is more suitable for high-accuracy monitoring of plasma etch responses obtained from OES.

Process operation improvement methodology based on statistical data analysis (통계적 분석기법을 이용한 공정 운전 향상의 방법)

  • Hwang, Dae-Hee;Ahn, Tae-Jin;Han, Chonghun
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1997.10a
    • /
    • pp.1516-1519
    • /
    • 1997
  • With disseminationof Distributed Control Systems(DCS), the huge amounts of process operation data could have been available and led to figure out process behaviors better on the statistical basis. Until now, the statistical modeling technology has been susally applied to process monitoring and fault diagnosis. however, it has been also thought that these process information, extracted from statistical analysis, might serve a great opportunity for process operation improvements and process improvements. This paper proposed a general methodolgy for process operation improvements including data analysis, backing up the result of analysis based on the methodology, and the mapping physical physical phenomena to the Principal Components(PC) which is the most distinguished feature in the methodology form traditional statistical analyses. The application of the proposed methodology to the Balst Furnace(BF) process has been presented for details. The BF process is one of the complicated processes, due to the highly nonlinear and correlated behaviors, and so the analysis for the process based on the mathematical modeling has been very difficult. So the statisitical analysis has come forward as a alternative way for the useful analysis. Using the proposed methodology, we could interpret the complicated process, the BF, better than any other mathematical methods and find the direction for process operation improvement. The direction of process operationimprovement, in the BF case, is to increase the fludization and the permeability, while decreasing the effect of tapping operation. These guide directions, with those physical meanings, could save fuel cost and process operator's pressure for proper actions, the better set point changes, in addition to the assistance with the better knowledge of the process. Open to set point change, the BF has a variety of steady state modes. In usual almost chemical processes are under the same situation with the BF in the point of multimode steady states. The proposed methodology focused on the application to the multimode steady state process such as the BF, consequently can be applied to any chemical processes set point changing whether operator intervened or not.

  • PDF