• Title/Summary/Keyword: Machine data analysis

Search Result 2,207, Processing Time 0.028 seconds

Finding the Optimal Data Classification Method Using LDA and QDA Discriminant Analysis

  • Kim, SeungJae;Kim, SungHwan
    • Journal of Integrative Natural Science
    • /
    • v.13 no.4
    • /
    • pp.132-140
    • /
    • 2020
  • With the recent introduction of artificial intelligence (AI) technology, the use of data is rapidly increasing, and newly generated data is also rapidly increasing. In order to obtain the results to be analyzed based on these data, the first thing to do is to classify the data well. However, when classifying data, if only one classification technique belonging to the machine learning technique is applied to classify and analyze it, an error of overfitting can be accompanied. In order to reduce or minimize the problems caused by misclassification of the classification system such as overfitting, it is necessary to derive an optimal classification by comparing the results of each classification by applying several classification techniques. If you try to interpret the data with only one classification technique, you will have poor reasoning and poor predictions of results. This study seeks to find a method for optimally classifying data by looking at data from various perspectives and applying various classification techniques such as LDA and QDA, such as linear or nonlinear classification, as a process before data analysis in data analysis. In order to obtain the reliability and sophistication of statistics as a result of big data analysis, it is necessary to analyze the meaning of each variable and the correlation between the variables. If the data is classified differently from the hypothesis test from the beginning, even if the analysis is performed well, unreliable results will be obtained. In other words, prior to big data analysis, it is necessary to ensure that data is well classified to suit the purpose of analysis. This is a process that must be performed before reaching the result by analyzing the data, and it may be a method of optimal data classification.

A Spatial Analysis of Seismic Vulnerability of Buildings Using Statistical and Machine Learning Techniques Comparative Analysis (통계분석 기법과 머신러닝 기법의 비교분석을 통한 건물의 지진취약도 공간분석)

  • Seong H. Kim;Sang-Bin Kim;Dae-Hyeon Kim
    • Journal of Industrial Convergence
    • /
    • v.21 no.1
    • /
    • pp.159-165
    • /
    • 2023
  • While the frequency of seismic occurrence has been increasing recently, the domestic seismic response system is weak, the objective of this research is to compare and analyze the seismic vulnerability of buildings using statistical analysis and machine learning techniques. As the result of using statistical technique, the prediction accuracy of the developed model through the optimal scaling method showed about 87%. As the result of using machine learning technique, because the accuracy of Random Forest method is 94% in case of Train Set, 76.7% in case of Test Set, which is the highest accuracy among the 4 analyzed methods, Random Forest method was finally chosen. Therefore, Random Forest method was derived as the final machine learning technique. Accordingly, the statistical analysis technique showed higher accuracy of about 87%, whereas the machine learning technique showed the accuracy of about 76.7%. As the final result, among the 22,296 analyzed building data, the seismic vulnerabilities of 1,627(0.1%) buildings are expected as more dangerous when the statistical analysis technique is used, 10,146(49%) buildings showed the same rate, and the remaining 10,523(50%) buildings are expected as more dangerous when the machine learning technique is used. As the comparison of the results of using advanced machine learning techniques in addition to the existing statistical analysis techniques, in spatial analysis decisions, it is hoped that this research results help to prepare more reliable seismic countermeasures.

VKOSPI Forecasting and Option Trading Application Using SVM (SVM을 이용한 VKOSPI 일 중 변화 예측과 실제 옵션 매매에의 적용)

  • Ra, Yun Seon;Choi, Heung Sik;Kim, Sun Woong
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.4
    • /
    • pp.177-192
    • /
    • 2016
  • Machine learning is a field of artificial intelligence. It refers to an area of computer science related to providing machines the ability to perform their own data analysis, decision making and forecasting. For example, one of the representative machine learning models is artificial neural network, which is a statistical learning algorithm inspired by the neural network structure of biology. In addition, there are other machine learning models such as decision tree model, naive bayes model and SVM(support vector machine) model. Among the machine learning models, we use SVM model in this study because it is mainly used for classification and regression analysis that fits well to our study. The core principle of SVM is to find a reasonable hyperplane that distinguishes different group in the data space. Given information about the data in any two groups, the SVM model judges to which group the new data belongs based on the hyperplane obtained from the given data set. Thus, the more the amount of meaningful data, the better the machine learning ability. In recent years, many financial experts have focused on machine learning, seeing the possibility of combining with machine learning and the financial field where vast amounts of financial data exist. Machine learning techniques have been proved to be powerful in describing the non-stationary and chaotic stock price dynamics. A lot of researches have been successfully conducted on forecasting of stock prices using machine learning algorithms. Recently, financial companies have begun to provide Robo-Advisor service, a compound word of Robot and Advisor, which can perform various financial tasks through advanced algorithms using rapidly changing huge amount of data. Robo-Adviser's main task is to advise the investors about the investor's personal investment propensity and to provide the service to manage the portfolio automatically. In this study, we propose a method of forecasting the Korean volatility index, VKOSPI, using the SVM model, which is one of the machine learning methods, and applying it to real option trading to increase the trading performance. VKOSPI is a measure of the future volatility of the KOSPI 200 index based on KOSPI 200 index option prices. VKOSPI is similar to the VIX index, which is based on S&P 500 option price in the United States. The Korea Exchange(KRX) calculates and announce the real-time VKOSPI index. VKOSPI is the same as the usual volatility and affects the option prices. The direction of VKOSPI and option prices show positive relation regardless of the option type (call and put options with various striking prices). If the volatility increases, all of the call and put option premium increases because the probability of the option's exercise possibility increases. The investor can know the rising value of the option price with respect to the volatility rising value in real time through Vega, a Black-Scholes's measurement index of an option's sensitivity to changes in the volatility. Therefore, accurate forecasting of VKOSPI movements is one of the important factors that can generate profit in option trading. In this study, we verified through real option data that the accurate forecast of VKOSPI is able to make a big profit in real option trading. To the best of our knowledge, there have been no studies on the idea of predicting the direction of VKOSPI based on machine learning and introducing the idea of applying it to actual option trading. In this study predicted daily VKOSPI changes through SVM model and then made intraday option strangle position, which gives profit as option prices reduce, only when VKOSPI is expected to decline during daytime. We analyzed the results and tested whether it is applicable to real option trading based on SVM's prediction. The results showed the prediction accuracy of VKOSPI was 57.83% on average, and the number of position entry times was 43.2 times, which is less than half of the benchmark (100 times). A small number of trading is an indicator of trading efficiency. In addition, the experiment proved that the trading performance was significantly higher than the benchmark.

Real-time Monitoring System for Rotating Machinery with IoT-based Cloud Platform (회전기계류 상태 실시간 진단을 위한 IoT 기반 클라우드 플랫폼 개발)

  • Jeong, Haedong;Kim, Suhyun;Woo, Sunhee;Kim, Songhyun;Lee, Seungchul
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.41 no.6
    • /
    • pp.517-524
    • /
    • 2017
  • The objective of this research is to improve the efficiency of data collection from many machine components on smart factory floors using IoT(Internet of things) techniques and cloud platform, and to make it easy to update outdated diagnostic schemes through online deployment methods from cloud resources. The short-term analysis is implemented by a micro-controller, and it includes machine-learning algorithms for inferring snapshot information of the machine components. For long-term analysis, time-series and high-dimension data are used for root cause analysis by combining a cloud platform and multivariate analysis techniques. The diagnostic results are visualized in a web-based display dashboard for an unconstrained user access. The implementation is demonstrated to identify its performance in data acquisition and analysis for rotating machinery.

A Study on the Reverse Engineering and Wear Analysis for Remanufacturing Planner Miller (플래너 밀러 재제조를 위한 역설계 및 마모 분석에 관한 연구)

  • Choi, Doo-Han;Kong, Seok-Hwan;Byeon, Jeong-Won;Kim, Tae-Woo;Hong, Dae-Sun
    • Journal of the Korean Society of Industry Convergence
    • /
    • v.25 no.6_2
    • /
    • pp.1103-1110
    • /
    • 2022
  • The old machine tools that have been used for a long time cause both increase in defective rate and decrease in productivity compared to new machines due to wear and failure of their components. In order to improve productivity and quality of machined components through remanufacturing, it is necessary to analyze the wear and failure of major components of old machine tools. In this study, the process for reverse engineering is designed for the remanufacture of planner millers, which belong to a very large machine tool. Also, the suitability of the designed process is verified through the analysis of the selected remanufactured components. In the first step of the process, some major components of the aging planner miller are scanned using a 3D laser scanner. In the next step, reverse engineering is performed using the data obtained through 3D scanning. Finally, wear and failure analysis is performed by comparing the reverse engineering data with the scan data. As a result, this reverse design and wear analysis can complement the insufficient design database and reduce costs in the maintenance of remanufactured products.

Optimal Design of an IPMSM for High-Speed Operation Using Electromagnetic and Stress Analysis

  • Seo, Jang-Ho;Jung, Hyun-Kyo
    • Journal of Electrical Engineering and Technology
    • /
    • v.4 no.3
    • /
    • pp.377-381
    • /
    • 2009
  • In the development of an interior permanent magnet synchronous machine (IPMSM) for high-speed operation, the problem of mechanical stress of the rotor by centrifugal force becomes more essential as the speed and size of the machines increase. In this paper, the optimal design process combined with mechanical stress analysis was presented. In the analysis of mechanical stress, the node and element data obtained by the electromagnetic field analysis program are also used in the stress analysis. Therefore, the different pre-processing for the stress analysis program is no longer required. Therefore, the computing time of the new method is very short compared with the conventional approach, and when repeated analyzes of various models are required, this method is very useful. The validity of our methods was verified by comparing simulation results with conventional and experimental data.

A Case Study on the Compatibility Analysis of Measurement Systems in Automobile Body Assembly

  • Lee, Myung-Duk;Lim, Ik-Sung;Sung, Chun-Ja
    • International Journal of Reliability and Applications
    • /
    • v.9 no.1
    • /
    • pp.7-15
    • /
    • 2008
  • The dimensional measurement equipment, such as Coordinate Measurement Machine (CMM), Optical Coordinate Measurement Machine (OCMM), and Checking Fixture (CF), take multiple dimensional measurements for each part in an automobile industry. Measurements are also recorded under different measurement systems to see if the responses differ significantly over these systems. Each measurement system (CMM, OCMM, and CF) will be considered as different treatments. This set-up provides massive amounts of process data which are multivariate in nature. Therefore, the multivariate statistical analysis is required to analyze data that are dependent on each other. This research provides step by step methodology for the evaluation procedure of the compatibility of measurement systems and clarify a systematic analyzation among the different measurement system's compatibility followed by number of case studies for each methodologies provided.

  • PDF

Sentiment Analysis to Evaluate Different Deep Learning Approaches

  • Sheikh Muhammad Saqib ;Tariq Naeem
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.11
    • /
    • pp.83-92
    • /
    • 2023
  • The majority of product users rely on the reviews that are posted on the appropriate website. Both users and the product's manufacturer could benefit from these reviews. Daily, thousands of reviews are submitted; how is it possible to read them all? Sentiment analysis has become a critical field of research as posting reviews become more and more common. Machine learning techniques that are supervised, unsupervised, and semi-supervised have worked very hard to harvest this data. The complicated and technological area of feature engineering falls within machine learning. Using deep learning, this tedious process may be completed automatically. Numerous studies have been conducted on deep learning models like LSTM, CNN, RNN, and GRU. Each model has employed a certain type of data, such as CNN for pictures and LSTM for language translation, etc. According to experimental results utilizing a publicly accessible dataset with reviews for all of the models, both positive and negative, and CNN, the best model for the dataset was identified in comparison to the other models, with an accuracy rate of 81%.

Prediction of Depression from Machine Learning Data (머신러닝 데이터의 우울증에 대한 예측)

  • Jeong Hee KIM;Kyung-A KIM
    • Journal of Korea Artificial Intelligence Association
    • /
    • v.1 no.1
    • /
    • pp.17-21
    • /
    • 2023
  • The primary objective of this research is to utilize machine learning models to analyze factors tailored to each dataset for predicting mental health conditions. The study aims to develop appropriate models based on specific datasets, with the goal of accurately predicting mental health states through the analysis of distinct factors present in each dataset. This approach seeks to design more effective strategies for the prevention and intervention of depression, enhancing the quality of mental health services by providing personalized services tailored to individual circumstances. Overall, the research endeavors to advance the development of personalized mental health prediction models through data-driven factor analysis, contributing to the improvement of mental health services on an individualized basis.

Analysis of Data Transfers in Java Virtual Machine (자바가상기계에서 데이터 이동 분석)

  • Yang, Hee-jae
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • v.9 no.1
    • /
    • pp.835-838
    • /
    • 2005
  • It is widely known that most operations performed in JVM belongs to data transfers at all times as JVM is based on abstract stack machine. Hence it is necessary to analyze the fashion of internal data transfers in JVM to develop a more efficient machine. We have analyzed in this paper the data transfer operations between operand stack, local variable array, heap, and constant pool in bytecode level.

  • PDF