• Title/Summary/Keyword: machine data

Search Result 6,279, Processing Time 0.035 seconds

Evaluation performance of machine learning in merging multiple satellite-based precipitation with gauge observation data

  • Nhuyen, Giang V.;Le, Xuan-hien;Jung, Sungho;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.143-143
    • /
    • 2022
  • Precipitation plays an essential role in water resources management and disaster prevention. Therefore, the understanding related to spatiotemporal characteristics of rainfall is necessary. Nowadays, highly accurate precipitation is mainly obtained from gauge observation systems. However, the density of gauge stations is a sparse and uneven distribution in mountainous areas. With the proliferation of technology, satellite-based precipitation sources are becoming increasingly common and can provide rainfall information in regions with complex topography. Nevertheless, satellite-based data is that it still remains uncertain. To overcome the above limitation, this study aims to take the strengthens of machine learning to generate a new reanalysis of precipitation data by fusion of multiple satellite precipitation products (SPPs) with gauge observation data. Several machine learning algorithms (i.e., Random Forest, Support Vector Regression, and Artificial Neural Network) have been adopted. To investigate the robustness of the new reanalysis product, observed data were collected to evaluate the accuracy of the products through Kling-Gupta efficiency (KGE), probability of detection (POD), false alarm rate (FAR), and critical success index (CSI). As a result, the new precipitation generated through the machine learning model showed higher accuracy than original satellite rainfall products, and its spatiotemporal variability was better reflected than others. Thus, reanalysis of satellite precipitation product based on machine learning can be useful source input data for hydrological simulations in ungauged river basins.

  • PDF

Genetic Algorithm Application to Machine Learning

  • Han, Myung-mook;Lee, Yill-byung
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.11 no.7
    • /
    • pp.633-640
    • /
    • 2001
  • In this paper we examine the machine learning issues raised by the domain of the Intrusion Detection Systems(IDS), which have difficulty successfully classifying intruders. There systems also require a significant amount of computational overhead making it difficult to create robust real-time IDS. Machine learning techniques can reduce the human effort required to build these systems and can improve their performance. Genetic algorithms are used to improve the performance of search problems, while data mining has been used for data analysis. Data Mining is the exploration and analysis of large quantities of data to discover meaningful patterns and rules. Among the tasks for data mining, we concentrate the classification task. Since classification is the basic element of human way of thinking, it is a well-studied problem in a wide variety of application. In this paper, we propose a classifier system based on genetic algorithm, and the proposed system is evaluated by applying it to IDS problem related to classification task in data mining. We report our experiments in using these method on KDD audit data.

  • PDF

Agent with Low-latency Overcoming Technique for Distributed Cluster-based Machine Learning

  • Seo-Yeon, Gu;Seok-Jae, Moon;Byung-Joon, Park
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.1
    • /
    • pp.157-163
    • /
    • 2023
  • Recently, as businesses and data types become more complex and diverse, efficient data analysis using machine learning is required. However, since communication in the cloud environment is greatly affected by network latency, data analysis is not smooth if information delay occurs. In this paper, SPT (Safe Proper Time) was applied to the cluster-based machine learning data analysis agent proposed in previous studies to solve this delay problem. SPT is a method of remotely and directly accessing memory to a cluster that processes data between layers, effectively improving data transfer speed and ensuring timeliness and reliability of data transfer.

Big Data Based Urban Transportation Analysis for Smart Cities - Machine Learning Based Traffic Prediction by Using Urban Environment Data - (도시 빅데이터를 활용한 스마트시티의 교통 예측 모델 - 환경 데이터와의 상관관계 기계 학습을 통한 예측 모델의 구축 및 검증 -)

  • Jang, Sun-Young;Shin, Dong-Youn
    • Journal of KIBIM
    • /
    • v.8 no.3
    • /
    • pp.12-19
    • /
    • 2018
  • The research aims to find implications of machine learning and urban big data as a way to construct the flexible transportation network system of smart city by responding the urban context changes. This research deals with a problem that existing a bus headway model is difficult to respond urban situations in real-time. Therefore, utilizing the urban big data and machine learning prototyping tool in weathers, traffics, and bus statues, this research presents a flexible headway model to predict bus delay and analyze the result. The prototyping model is composed by real-time data of buses. The data is gathered through public data portals and real time Application Program Interface (API) by the government. These data are fundamental resources to organize interval pattern models of bus operations as traffic environment factors (road speeds, station conditions, weathers, and bus information of operating in real-time). The prototyping model is implemented by the machine learning tool (RapidMiner Studio) and conducted several tests for bus delays prediction according to specific circumstances. As a result, possibilities of transportation system are discussed for promoting the urban efficiency and the citizens' convenience by responding to urban conditions.

A Study on Error Detection Algorithm of COD Measurement Machine

  • Choi, Hyun-Seok;Song, Gyu-Moon;Kim, Tae-Yoon
    • Journal of the Korean Data and Information Science Society
    • /
    • v.18 no.4
    • /
    • pp.847-857
    • /
    • 2007
  • This paper provides a statistical algorithm which detects COD (chemical oxygen demand) measurement machine error on real-time. For this we propose to use regression model fitting and check its validity against the current observations. The main idea is that the normal regression relation between COD measurement and other parameters inside the machine will be violated when the machine is out of order.

  • PDF

Data Security on Cloud by Cryptographic Methods Using Machine Learning Techniques

  • Gadde, Swetha;Amutharaj, J.;Usha, S.
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.5
    • /
    • pp.342-347
    • /
    • 2022
  • On Cloud, the important data of the user that is protected on remote servers can be accessed via internet. Due to rapid shift in technology nowadays, there is a swift increase in the confidential and pivotal data. This comes up with the requirement of data security of the user's data. Data is of different type and each need discrete degree of conservation. The idea of data security data science permits building the computing procedure more applicable and bright as compared to conventional ones in the estate of data security. Our focus with this paper is to enhance the safety of data on the cloud and also to obliterate the problems associated with the data security. In our suggested plan, some basic solutions of security like cryptographic techniques and authentication are allotted in cloud computing world. This paper put your heads together about how machine learning techniques is used in data security in both offensive and defensive ventures, including analysis on cyber-attacks focused at machine learning techniques. The machine learning technique is based on the Supervised, UnSupervised, Semi-Supervised and Reinforcement Learning. Although numerous research has been done on this topic but in reference with the future scope a lot more investigation is required to be carried out in this field to determine how the data can be secured more firmly on cloud in respect with the Machine Learning Techniques and cryptographic methods.

A Study on Machine Failure Improvement Using F-RPN(Failure-RPN): Focusing on the Semiconductor Etching Process (F-RPN(Failure-RPN)을 이용한 장비 고장률 개선 연구: 반도체 식각 공정을 중심으로)

  • Lee, Hyung-Geun;Hong, Yong-Min;Kang, Sung-Woo
    • Journal of the Korea Safety Management & Science
    • /
    • v.23 no.3
    • /
    • pp.27-33
    • /
    • 2021
  • The purpose of this study is to present a novel indicator for analyzing machine failure based on its idle time and productivity. Existing machine repair plan was limited to machine experts from its manufacturing industries. This study evaluates the repair status of machines and extracts machines that need improvement. In this study, F-RPN was calculated using the etching process data provided by the 2018 PHM Data Challenge. Each S(S: Severity), O(O: Occurence), D(D: Detection) is divided into the idle time of the machine, the number of fault data, and the failure rate, respectively. The repair status of machine is quantified through the F-RPN calculated by multiplying S, O, and D. This study conducts a case study of machine in a semiconductor etching process. The process capability index has the disadvantage of not being able to divide the values outside the range. The performance of this index declines when the manufacturing process is under control, hereby introducing F-RPN to evaluate machine status that are difficult to distinguish by process capability index.

Real-time 3D Graphic Simulation of the Spent Fuel Rod Extracting Machine for Remote Monitoring (사용후핵연료봉 인출장치의 원격감시를 위한 실시간 3차원 그래픽 시뮬레이션)

  • 송태길;이종열;김성현;윤지섭
    • Korean Journal of Computational Design and Engineering
    • /
    • v.5 no.4
    • /
    • pp.327-335
    • /
    • 2000
  • The spent fuel rod extracting machine is automatically operated in high radioactive environment, so high reliability of operation is required. In this paper, to enhance the reliability of this machine by providing a close monitoring capability. a real time graphic simulation method is suggested. This method utilizes conventional IGRIP (Interactive Graphics Robot Instruction Program) 3D graphic simulation tool to visualize and simulate the 3D graphic model of this machine. Also, the dedicated protocol is defined for transmission of the operational data of the machine. The real time graphic simulation is realized by developing the socket module between a graphic workstation and a machine control computer through the TCP/IP network and by dividing the 3D graphic simulation GSL(Graphic Simulation Language) program as a small sized sub routine. The suggested method is implemented while automatically operating the rod extracting machine. The result of implementation shows that the real time 3D graphic simulation is well synchronized with the actual machine according to the operational data.

  • PDF

Quadratic Loss Support Vector Interval Regression Machine for Crisp Input-Output Data

  • Hwang, Chang-Ha
    • Journal of the Korean Data and Information Science Society
    • /
    • v.15 no.2
    • /
    • pp.449-455
    • /
    • 2004
  • Support vector machine (SVM) has been very successful in pattern recognition and function estimation problems for crisp data. This paper proposes a new method to evaluate interval regression models for crisp input-output data. The proposed method is based on quadratic loss SVM, which implements quadratic programming approach giving more diverse spread coefficients than a linear programming one. The proposed algorithm here is model-free method in the sense that we do not have to assume the underlying model function. Experimental result is then presented which indicate the performance of this algorithm.

  • PDF

Design of a Node Label Data Flow Machine based on Self-timed (Self-timed 기반의 Node Label Data Flow Machine 설계)

  • Kim, Hee-Sook;Jung, Sung-Tae;Park, Hee-Soon
    • Proceedings of the KIEE Conference
    • /
    • 1998.11b
    • /
    • pp.666-668
    • /
    • 1998
  • In this paper we illustrate the design of a node label data flow machine based on self-timed paradigm. Data flow machines differ from most other parallel architectures, they are based on the concept of the data-driven computation model instead of the program store computation model. Since the data-driven computation model provides the excution of instructions asynchronously, it is natural to implement a data flow machine using self timed circuits.

  • PDF