• Title/Summary/Keyword: analysis data

Search Result 85,083, Processing Time 0.082 seconds

Study of Analysis Software for Event Recorder in High Speed Railway (고속전철용 Event Recorder를 위한 분석도구 소프트웨어 연구)

  • Song, Gyu-Youn;Lee, Sang-Nam;Ryu, Hee-Moon;Kim, Kwang-Yul;Han, Kwang-Rok
    • Proceedings of the KSR Conference
    • /
    • 2009.05b
    • /
    • pp.341-347
    • /
    • 2009
  • In high speed railway, event recorder system stores a train speed and the related data for train operation in real time. Using those information, we can analysis the train operation and the reason of train accident. Analysis software gets the stored data from Event Recorder and shows the status of various signals related with train operation. Using it, also we can analysis the train operation before and after the given time. In this paper we propose the analysis software to show and analysis the operation of high speed train. The method of transferring the stored data from Event Recorder into Analysis Software is proposed. We develop the efficient procedure to store the transferred data into analysis system. Also the effective method to show the store data and to analysis them is studied for finding the cause of train accident.

  • PDF

Keyword Analysis of Data Technology Using Big Data Technique (빅데이터 기법을 활용한 Data Technology의 키워드 분석)

  • Park, Sung-Uk
    • Journal of Korea Technology Innovation Society
    • /
    • v.22 no.2
    • /
    • pp.265-281
    • /
    • 2019
  • With the advent of the Internet-based economy, the dramatic changes in consumption patterns have been witnessed during the last decades. The seminal change has led by Data Technology, the integrated platform of mobile, online, offline and artificial intelligence, which remained unchallenged. In this paper, I use data analysis tool (TexTom) in order to articulate the definitfite notion of data technology from Internet sources. The data source is collected for last three years (November 2015 ~ November 2018) from Google and Naver. And I have derived several key keywords related to 'Data Technology'. As a result, it was found that the key keyword technologies of Big Data, O2O (Offline-to-Online), AI, IoT (Internet of things), and cloud computing are related to Data Technology. The results of this study can be used as useful information that can be referred to when the Data Technology age comes.

Development of Web-based Off-site Consequence Analysis Program and its Application for ILRT Extension (격납건물종합누설률시험 주기연장을 위한 웹기반 소외결말분석 프로그램 개발 및 적용)

  • Na, Jang-Hwan;Hwang, Seok-Won;Oh, Ji-Yong
    • Journal of the Korean Society of Safety
    • /
    • v.27 no.5
    • /
    • pp.219-223
    • /
    • 2012
  • For an off-site consequence analysis at nuclear power plant, MELCOR Accident Consequence Code System(MACCS) II code is widely used as a software tool. In this study, the algorithm of web-based off-site consequence analysis program(OSCAP) using the MACCS II code was developed for an Integrated Leak Rate Test (ILRT) interval extension and Level 3 probabilistic safety assessment(PSA), and verification and validation(V&V) of the program was performed. The main input data for the MACCS II code are meteorological, population distribution and source term information. However, it requires lots of time and efforts to generate the main input data for an off-site consequence analysis using the MACCS II code. For example, the meteorological data are collected from each nuclear power site in real time, but the formats of the raw data collected are different from each site. To reduce the efforts and time for risk assessments, the web-based OSCAP has an automatic processing module which converts the format of the raw data collected from each site to the input data format of the MACCS II code. The program also provides an automatic function of converting the latest population data from Statistics Korea, the National Statistical Office, to the population distribution input data format of the MACCS II code. For the source term data, the program includes the release fraction of each source term category resulting from modular accident analysis program(MAAP) code analysis and the core inventory data from ORIGEN. These analysis results of each plant in Korea are stored in a database module of the web-based OSCAP, so the user can select the defaulted source term data of each plant without handling source term input data.

A Study of Choice for Analysis Method on Repeated Measures Clinical Data

  • Song, Jung
    • Korean Journal of Clinical Laboratory Science
    • /
    • v.45 no.2
    • /
    • pp.60-65
    • /
    • 2013
  • Data from repeated measurements are accomplished through repeatedly processing the same subject under different conditions and different points of view. The power of testing enhances the choice of pertinent analysis methods that agrees with the characteristics of data concerned and the situation involved. Along with the clinical example, this paper compares the analysis of the variance on ex-post tests, gain score analysis, analysis by mixed design and analysis of covariance employable for repeating measure. Comparing the analysis of variance on ex post test, and gain score analysis on correlations, leads to the fact that the latter enhances the power of the test and diminishes the variance of error terms. The concluded probability, identified that the gain score analysis and the mixed design on interaction between "between subjects factor" and "within subjects factor", are identical. The analysis of covariance, demonstrated better power of the test and smaller error terms than the gain score analysis. Research on four analysis method found that the analysis of covariance is the most appropriate in clinical data than two repeated test with high correlation and ex ante affects ex post.

  • PDF

Data anomaly detection and Data fusion based on Incremental Principal Component Analysis in Fog Computing

  • Yu, Xue-Yong;Guo, Xin-Hui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.10
    • /
    • pp.3989-4006
    • /
    • 2020
  • The intelligent agriculture monitoring is based on the perception and analysis of environmental data, which enables the monitoring of the production environment and the control of environmental regulation equipment. As the scale of the application continues to expand, a large amount of data will be generated from the perception layer and uploaded to the cloud service, which will bring challenges of insufficient bandwidth and processing capacity. A fog-based offline and real-time hybrid data analysis architecture was proposed in this paper, which combines offline and real-time analysis to enable real-time data processing on resource-constrained IoT devices. Furthermore, we propose a data process-ing algorithm based on the incremental principal component analysis, which can achieve data dimensionality reduction and update of principal components. We also introduce the concept of Squared Prediction Error (SPE) value and realize the abnormal detection of data through the combination of SPE value and data fusion algorithm. To ensure the accuracy and effectiveness of the algorithm, we design a regular-SPE hybrid model update strategy, which enables the principal component to be updated on demand when data anomalies are found. In addition, this strategy can significantly reduce resource consumption growth due to the data analysis architectures. Practical datasets-based simulations have confirmed that the proposed algorithm can perform data fusion and exception processing in real-time on resource-constrained devices; Our model update strategy can reduce the overall system resource consumption while ensuring the accuracy of the algorithm.

Curriculum of Basic Data Science Practices for Non-majors (비전공자 대상 기초 데이터과학 실습 커리큘럼)

  • Hur, Kyeong
    • Journal of Practical Engineering Education
    • /
    • v.12 no.2
    • /
    • pp.265-273
    • /
    • 2020
  • In this paper, to design a basic data science practice curriculum as a liberal arts subject for non-majors, we proposed an educational method using an Excel(spreadsheet) data analysis tool. Tools for data collection, data processing, and data analysis include Excel, R, Python, and Structured Query Language (SQL). When it comes to practicing data science, R, Python and SQL need to understand programming languages and data structures together. On the other hand, the Excel tool is a data analysis tool familiar to the general public, and it does not have the burden of learning a programming language. And if you practice basic data science practice with Excel, you have the advantage of being able to concentrate on acquiring data science content. In this paper, a basic data science practice curriculum for one semester and weekly Excel practice contents were proposed. And, to demonstrate the substance of the educational content, examples of Linear Regression Analysis were presented using Excel data analysis tools.

An Analysis of Domestic Research Trend on Research Data Using Keyword Network Analysis (키워드 네트워크 분석을 이용한 연구데이터 관련 국내 연구 동향 분석)

  • Sangwoo Han
    • Journal of Korean Library and Information Science Society
    • /
    • v.54 no.4
    • /
    • pp.393-414
    • /
    • 2023
  • The goal of this study is to investigate domestic research trend on research data study. To achieve this goal, articles related research data topic were collected from RISS. After data cleansing, 134 author keywords were extracted from a total of 58 articles and keyword network analysis was performed. As a result, first, the number of studies related to research data in Korea is still only 58, so it was found that many related studies need to be conducted in the future. Second, most research fields related to research data were focused on library and information science among complex studies. Third, as a result of frequency analysis of author keywords related to research data, 'research data management', 'research data sharing', 'data repository', and 'open science' were analyzed as major frequent keywords, so research data-related research focuses on the above keywords. The keyword network analysis results also showed that high-frequency keywords occupy a central position in degree centrality and betweenness centrality and are located as core keywords in related studies. Through the results of this study, we were able to identify trends related to recent research data and identify areas that require intensive research in the future.

Estimation of Material Requirement of Piping Materials in an Offshore Structure using Big Data Analysis (빅데이터 분석을 이용한 해양 구조물 배관 자재의 소요량 예측)

  • Oh, Min-Jae;Roh, Myung-Il;Park, Sung-Woo;Kim, Seong-Hoon
    • Journal of the Society of Naval Architects of Korea
    • /
    • v.55 no.3
    • /
    • pp.243-251
    • /
    • 2018
  • In the shipyard, a lot of data is generated, stored, and managed during design, construction, and operation phases to build ships and offshore structures. However, it is difficult to handle such big data efficiently using existing data-handling technologies. As the big data technology is developed, the ship and offshore industries start to focus on the existing big data to find valuable information from it. In this paper, the material requirement estimation method of offshore structure piping materials using big data analysis is proposed. A big data platform for the data analysis in the shipyard is introduced and it is applied to the analysis of material requirement estimation to solve the problems in piping design by a designer. The regression model is developed from the big data of piping materials and verified using the existing data. This analysis can help a piping designer to estimate the exact amount of material requirement and schedule the purchase time.

The effect of missing levels of nesting in multilevel analysis

  • Park, Seho;Chung, Yujin
    • Genomics & Informatics
    • /
    • v.20 no.3
    • /
    • pp.34.1-34.11
    • /
    • 2022
  • Multilevel analysis is an appropriate and powerful tool for analyzing hierarchical structure data widely applied from public health to genomic data. In practice, however, we may lose the information on multiple nesting levels in the multilevel analysis since data may fail to capture all levels of hierarchy, or the top or intermediate levels of hierarchy are ignored in the analysis. In this study, we consider a multilevel linear mixed effect model (LMM) with single imputation that can involve all data hierarchy levels in the presence of missing top or intermediate-level clusters. We evaluate and compare the performance of a multilevel LMM with single imputation with other models ignoring the data hierarchy or missing intermediate-level clusters. To this end, we applied a multilevel LMM with single imputation and other models to hierarchically structured cohort data with some intermediate levels missing and to simulated data with various cluster sizes and missing rates of intermediate-level clusters. A thorough simulation study demonstrated that an LMM with single imputation estimates fixed coefficients and variance components of a multilevel model more accurately than other models ignoring data hierarchy or missing clusters in terms of mean squared error and coverage probability. In particular, when models ignoring data hierarchy or missing clusters were applied, the variance components of random effects were overestimated. We observed similar results from the analysis of hierarchically structured cohort data.

A Study on the Analysis Techniques for Big Data Computing (빅데이터 컴퓨팅을 위한 분석기법에 관한 연구)

  • Oh, Sun-Jin
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.3
    • /
    • pp.475-480
    • /
    • 2021
  • With the rapid development of mobile, cloud computing technology and social network services, we are in the flood of huge data and realize that these large-scale data contain very precious value and important information. Big data, however, have both latent useful value and critical risks, so, nowadays, a lot of researches and applications for big data has been executed actively in order to extract useful information from big data efficiently and make the most of the potential information effectively. At this moment, the data analysis technique that can extract precious information from big data efficiently is the most important step in big data computing process. In this study, we investigate various data analysis techniques that can extract the most useful information in big data computing process efficiently, compare pros and cons of those techniques, and propose proper data analysis method that can help us to find out the best solution of the big data analysis in the peculiar situation.