• Title/Summary/Keyword: methods of data analysis

Search Result 19,359, Processing Time 0.051 seconds

Experimental validation of ASME strain-based seismic assessment methods using piping elbow test data

  • Jong-Min Lee ;Jae-Yoon Kim;Hyun-Seok Song ;Yun-Jae Kim ;Jin-Weon Kim
    • Nuclear Engineering and Technology
    • /
    • v.55 no.5
    • /
    • pp.1616-1629
    • /
    • 2023
  • To quantify the conservatism of existing ASME strain-based evaluation methods for seismic loading, this paper presents very low cycle fatigue test data of elbows under various cyclic loading conditions and comparison of evaluation results with experimental failure cycles. For strain-based evaluation methods, the method presented in ASME BPVC CC N-900 and Sec. VIII are used. Predicted failure cycles are compared with experimental failure cycle to quantify the conservatism of evaluation methods. All methods give very conservative failure cycles. The CC N-900 method is the most conservative and prediction results are only ~0.5% of experimental data. For Sec. VIII method, the use of the option using code tensile properties gives ~3% of experimental data, and the use of the material-specific reduction of area can reduce conservatism but still gives ~15% of experimental data.

Analysis of somatotype using korean elderly women's anthropometric measurement

  • Yoo, B.C.;Park, I.S.;Kim, Y.S.
    • Journal of the Ergonomics Society of Korea
    • /
    • v.15 no.2
    • /
    • pp.149-155
    • /
    • 1996
  • The purpose of this study is to provide basic information for elderly women's clothing design and to develop dress forms that can reflect the characteristics of their bodies. Three hundered twenty subjects, between the ages of 60 .approx. 85, were chosen and the data were collected from anthropometric data. Principal component analysis which is one of the basic methods in factor analysis was applied to the interpretation of anthropometric data. As a result of analysis, the data are able to be decided into appropriate group.

  • PDF

Automatic Cleaning Algorithm of Asset Data for Transmission Cable (지중 송전케이블 자산데이터의 자동 정제 알고리즘 개발연구)

  • Hwang, Jae-Sang;Mun, Sung-Duk;Kim, Tae-Joon;Kim, Kang-Sik
    • KEPCO Journal on Electric Power and Energy
    • /
    • v.7 no.1
    • /
    • pp.79-84
    • /
    • 2021
  • The fundamental element to be kept for big data analysis, artificial intelligence technologies and asset management system is a data quality, which could directly affect the entire system reliability. For this reason, the momentum of data cleaning works is recently increased and data cleaning methods have been investigating around the world. In the field of electric power, however, asset data cleaning methods have not been fully determined therefore, automatic cleaning algorithm of asset data for transmission cables has been studied in this paper. Cleaning algorithm is composed of missing data treatment and outlier data one. Rule-based and expert opinion based cleaning methods are converged and utilized for these dirty data.

A Method for Microarray Data Analysis based on Bayesian Networks using an Efficient Structural learning Algorithm and Data Dimensionality Reduction (효율적 구조 학습 알고리즘과 데이타 차원축소를 통한 베이지안망 기반의 마이크로어레이 데이타 분석법)

  • 황규백;장정호;장병탁
    • Journal of KIISE:Software and Applications
    • /
    • v.29 no.11
    • /
    • pp.775-784
    • /
    • 2002
  • Microarray data, obtained from DNA chip technologies, is the measurement of the expression level of thousands of genes in cells or tissues. It is used for gene function prediction or cancer diagnosis based on gene expression patterns. Among diverse methods for data analysis, the Bayesian network represents the relationships among data attributes in the form of a graph structure. This property enables us to discover various relations among genes and the characteristics of the tissue (e.g., the cancer type) through microarray data analysis. However, most of the present microarray data sets are so sparse that it is difficult to apply general analysis methods, including Bayesian networks, directly. In this paper, we harness an efficient structural learning algorithm and data dimensionality reduction in order to analyze microarray data using Bayesian networks. The proposed method was applied to the analysis of real microarray data, i.e., the NC160 data set. And its usefulness was evaluated based on the accuracy of the teamed Bayesian networks on representing the known biological facts.

A Study on Improving the predict accuracy rate of Hybrid Model Technique Using Error Pattern Modeling : Using Logistic Regression and Discriminant Analysis

  • Cho, Yong-Jun;Hur, Joon
    • Journal of the Korean Data and Information Science Society
    • /
    • v.17 no.2
    • /
    • pp.269-278
    • /
    • 2006
  • This paper presents the new hybrid data mining technique using error pattern, modeling of improving classification accuracy. The proposed method improves classification accuracy by combining two different supervised learning methods. The main algorithm generates error pattern modeling between the two supervised learning methods(ex: Neural Networks, Decision Tree, Logistic Regression and so on.) The Proposed modeling method has been applied to the simulation of 10,000 data sets generated by Normal and exponential random distribution. The simulation results show that the performance of proposed method is superior to the existing methods like Logistic regression and Discriminant analysis.

  • PDF

Network-based Microarray Data Analysis Tool

  • Park, Hee-Chang;Ryu, Ki-Hyun
    • Journal of the Korean Data and Information Science Society
    • /
    • v.17 no.1
    • /
    • pp.53-62
    • /
    • 2006
  • DNA microarray data analysis is a new technology to investigate the expression levels of thousands of genes simultaneously. Since DNA microarray data structures are various and complicative, the data are generally stored in databases for approaching to and controlling the data effectively. But we have some difficulties to analyze and control the data when the data are stored in the several database management systems or that the data are stored to the file format. The existing analysis tools for DNA microarray data have many difficult problems by complicated instructions, and dependency on data types and operating system. In this paper, we design and implement network-based analysis tool for obtaining to useful information from DNA microarray data. When we use this tool, we can analyze effectively DNA microarray data without special knowledge and education for data types and analytical methods.

  • PDF

Techniques of Selection of Landslide Hazard Area on Analysis of Topographical Data (지형 데이타 해석에 따른 산사태 위험지역 선정 기법)

  • 강인준;최철웅;장용구
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.12 no.2
    • /
    • pp.147-154
    • /
    • 1994
  • Landslides in the close residential area occur loss of life and properties. Usually soil mechanical methods and multi-statistical methods used to analysis the landslide. The factors of landslide analysis by multi-statistical methods are composite operations of geology, topography, vegetation, and climate etc. In this study, existing landslide areas are cheesed to extract the characteristics of their geographical data for adaption of Multi-statical methods. Authors give a weight to their values and analyze the some area using technique of overlap on GSIS.

  • PDF

A Universal Analysis Pipeline for Hybrid Capture-Based Targeted Sequencing Data with Unique Molecular Indexes

  • Kim, Min-Jung;Kim, Si-Cho;Kim, Young-Joon
    • Genomics & Informatics
    • /
    • v.16 no.4
    • /
    • pp.29.1-29.5
    • /
    • 2018
  • Hybrid capture-based targeted sequencing is being used increasingly for genomic variant profiling in tumor patients. Unique molecular index (UMI) technology has recently been developed and helps to increase the accuracy of variant calling by minimizing polymerase chain reaction biases and sequencing errors. However, UMI-adopted targeted sequencing data analysis is slightly different from the methods for other types of omics data, and its pipeline for variant calling is still being optimized in various study groups for their own purposes. Due to this provincial usage of tools, our group built an analysis pipeline for global application to many studies of targeted sequencing generated with different methods. First, we generated hybrid capture-based data using genomic DNA extracted from tumor tissues of colorectal cancer patients. Sequencing libraries were prepared and pooled together, and an 8-plexed capture library was processed to the enrichment step before 150-bp paired-end sequencing with Illumina HiSeq series. For the analysis, we evaluated several published tools. We focused mainly on the compatibility of the input and output of each tool. Finally, our laboratory built an analysis pipeline specialized for UMI-adopted data. Through this pipeline, we were able to estimate even on-target rates and filtered consensus reads for more accurate variant calling. These results suggest the potential of our analysis pipeline in the precise examination of the quality and efficiency of conducted experiments.

Landslide Susceptibility Analysis and its Verification using Likelihood Ratio, Logistic Regression and Artificial Neural Network Methods: Case study of Yongin, Korea

  • Lee, S.;Ryu, J. H.
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.132-134
    • /
    • 2003
  • The likelihood ratio, logistic regression and artificial neural networks methods are applied and verified for analysis of landslide susceptibility in Yongin, Korea using GIS. From a spatial database containing such data as landslide location, topography, soil, forest, geology and land use, the 14 landsliderelated factors were calculated or extracted. Using these factors, landslide susceptibility indexes were calculated by likelihood ratio, logistic regression and artificial neural network methods. Before the calculation, the study area was divided into two sides (west and east) of equal area, for verification of the methods. Thus, the west side was used to assess the landslide susceptibility, and the east side was used to verify the derived susceptibility. The results of the landslide susceptibility analysis were verified using success and prediction rates. The v erification results showed satisfactory agreement between the susceptibility map and the exis ting data on landslide locations.

  • PDF

A study on the properties of sensitivity analysis in principal component regression and latent root regression (주성분회귀와 고유값회귀에 대한 감도분석의 성질에 대한 연구)

  • Shin, Jae-Kyoung;Chang, Duk-Joon
    • Journal of the Korean Data and Information Science Society
    • /
    • v.20 no.2
    • /
    • pp.321-328
    • /
    • 2009
  • In regression analysis, the ordinary least squares estimates of regression coefficients become poor, when the correlations among predictor variables are high. This phenomenon, which is called multicollinearity, causes serious problems in actual data analysis. To overcome this multicollinearity, many methods have been proposed. Ridge regression, shrinkage estimators and methods based on principal component analysis (PCA) such as principal component regression (PCR) and latent root regression (LRR). In the last decade, many statisticians discussed sensitivity analysis (SA) in ordinary multiple regression and same topic in PCR, LRR and logistic principal component regression (LPCR). In those methods PCA plays important role. Many statisticians discussed SA in PCA and related multivariate methods. We introduce the method of PCR and LRR. We also introduce the methods of SA in PCR and LRR, and discuss the properties of SA in PCR and LRR.

  • PDF