• Title/Summary/Keyword: Learning Performance Prediction

Search Result 1,122, Processing Time 0.033 seconds

Development of Prediction Model of Chloride Diffusion Coefficient using Machine Learning (기계학습을 이용한 염화물 확산계수 예측모델 개발)

  • Kim, Hyun-Su
    • Journal of Korean Association for Spatial Structures
    • /
    • v.23 no.3
    • /
    • pp.87-94
    • /
    • 2023
  • Chloride is one of the most common threats to reinforced concrete (RC) durability. Alkaline environment of concrete makes a passive layer on the surface of reinforcement bars that prevents the bar from corrosion. However, when the chloride concentration amount at the reinforcement bar reaches a certain level, deterioration of the passive protection layer occurs, causing corrosion and ultimately reducing the structure's safety and durability. Therefore, understanding the chloride diffusion and its prediction are important to evaluate the safety and durability of RC structure. In this study, the chloride diffusion coefficient is predicted by machine learning techniques. Various machine learning techniques such as multiple linear regression, decision tree, random forest, support vector machine, artificial neural networks, extreme gradient boosting annd k-nearest neighbor were used and accuracy of there models were compared. In order to evaluate the accuracy, root mean square error (RMSE), mean square error (MSE), mean absolute error (MAE) and coefficient of determination (R2) were used as prediction performance indices. The k-fold cross-validation procedure was used to estimate the performance of machine learning models when making predictions on data not used during training. Grid search was applied to hyperparameter optimization. It has been shown from numerical simulation that ensemble learning methods such as random forest and extreme gradient boosting successfully predicted the chloride diffusion coefficient and artificial neural networks also provided accurate result.

Extrapolation of wind pressure for low-rise buildings at different scales using few-shot learning

  • Yanmo Weng;Stephanie G. Paal
    • Wind and Structures
    • /
    • v.36 no.6
    • /
    • pp.367-377
    • /
    • 2023
  • This study proposes a few-shot learning model for extrapolating the wind pressure of scaled experiments to full-scale measurements. The proposed ML model can use scaled experimental data and a few full-scale tests to accurately predict the remaining full-scale data points (for new specimens). This model focuses on extrapolating the prediction to different scales while existing approaches are not capable of accurately extrapolating from scaled data to full-scale data in the wind engineering domain. Also, the scaling issue observed in wind tunnel tests can be partially resolved via the proposed approach. The proposed model obtained a low mean-squared error and a high coefficient of determination for the mean and standard deviation wind pressure coefficients of the full-scale dataset. A parametric study is carried out to investigate the influence of the number of selected shots. This technique is the first of its kind as it is the first time an ML model has been used in the wind engineering field to deal with extrapolation in wind performance prediction. With the advantages of the few-shot learning model, physical wind tunnel experiments can be reduced to a great extent. The few-shot learning model yields a robust, efficient, and accurate alternative to extrapolating the prediction performance of structures from various model scales to full-scale.

Effects of CNN Backbone on Trajectory Prediction Models for Autonomous Vehicle

  • Seoyoung Lee;Hyogyeong Park;Yeonhwi You;Sungjung Yong;Il-Young Moon
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.4
    • /
    • pp.346-350
    • /
    • 2023
  • Trajectory prediction is an essential element for driving autonomous vehicles, and various trajectory prediction models have emerged with the development of deep learning technology. Convolutional neural network (CNN) is the most commonly used neural network architecture for extracting the features of visual images, and the latest models exhibit high performances. This study was conducted to identify an efficient CNN backbone model among the components of deep learning models for trajectory prediction. We changed the existing CNN backbone network of multiple-trajectory prediction models used as feature extractors to various state-of-the-art CNN models. The experiment was conducted using nuScenes, which is a dataset used for the development of autonomous vehicles. The results of each model were compared using frequently used evaluation metrics for trajectory prediction. Analyzing the impact of the backbone can improve the performance of the trajectory prediction task. Investigating the influence of the backbone on multiple deep learning models can be a future challenge.

Support Vector Bankruptcy Prediction Model with Optimal Choice of RBF Kernel Parameter Values using Grid Search (Support Vector Machine을 이용한 부도예측모형의 개발 -격자탐색을 이용한 커널 함수의 최적 모수 값 선정과 기존 부도예측모형과의 성과 비교-)

  • Min Jae H.;Lee Young-Chan
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.30 no.1
    • /
    • pp.55-74
    • /
    • 2005
  • Bankruptcy prediction has drawn a lot of research interests in previous literature, and recent studies have shown that machine learning techniques achieved better performance than traditional statistical ones. This paper employs a relatively new machine learning technique, support vector machines (SVMs). to bankruptcy prediction problem in an attempt to suggest a new model with better explanatory power and stability. To serve this purpose, we use grid search technique using 5-fold cross-validation to find out the optimal values of the parameters of kernel function of SVM. In addition, to evaluate the prediction accuracy of SVM. we compare its performance with multiple discriminant analysis (MDA), logistic regression analysis (Logit), and three-layer fully connected back-propagation neural networks (BPNs). The experiment results show that SVM outperforms the other methods.

Early Software Quality Prediction Using Support Vector Machine (Support Vector Machine을 이용한 초기 소프트웨어 품질 예측)

  • Hong, Euy-Seok
    • Journal of Information Technology Services
    • /
    • v.10 no.2
    • /
    • pp.235-245
    • /
    • 2011
  • Early criticality prediction models that determine whether a design entity is fault-prone or not are becoming more and more important as software development projects are getting larger. Effective predictions can reduce the system development cost and improve software quality by identifying trouble-spots at early phases and proper allocation of effort and resources. Many prediction models have been proposed using statistical and machine learning methods. This paper builds a prediction model using Support Vector Machine(SVM) which is one of the most popular modern classification methods and compares its prediction performance with a well-known prediction model, BackPropagation neural network Model(BPM). SVM is known to generalize well even in high dimensional spaces under small training data conditions. In prediction performance evaluation experiments, dimensionality reduction techniques for data set are not used because the dimension of input data is too small. Experimental results show that the prediction performance of SVM model is slightly better than that of BPM and polynomial kernel function achieves better performance than other SVM kernel functions.

Comparison of Wave Prediction and Performance Evaluation in Korea Waters based on Machine Learning

  • Heung Jin Park;Youn Joung Kang
    • Journal of Ocean Engineering and Technology
    • /
    • v.38 no.1
    • /
    • pp.18-29
    • /
    • 2024
  • Waves are a complex phenomenon in marine and coastal areas, and accurate wave prediction is essential for the safety and resource management of ships at sea. In this study, three types of machine learning techniques specialized in nonlinear data processing were used to predict the waves of Korea waters. An optimized algorithm for each area is presented for performance evaluation and comparison. The optimal parameters were determined by varying the window size, and the performance was evaluated by comparing the mean absolute error (MAE). All the models showed good results when the window size was 4 or 7 d, with the gated recurrent unit (GRU) performing well in all waters. The MAE results were within 0.161 m to 0.051 m for significant wave heights and 0.491 s to 0.272 s for periods. In addition, the GRU showed higher prediction accuracy for certain data with waves greater than 3 m or 8 s, which is likely due to the number of training parameters. When conducting marine and offshore research at new locations, the results presented in this study can help ensure safety and improve work efficiency. If additional wave-related data are obtained, more accurate wave predictions will be possible.

The Effect of Input Variables Clustering on the Characteristics of Ensemble Machine Learning Model for Water Quality Prediction (입력자료 군집화에 따른 앙상블 머신러닝 모형의 수질예측 특성 연구)

  • Park, Jungsu
    • Journal of Korean Society on Water Environment
    • /
    • v.37 no.5
    • /
    • pp.335-343
    • /
    • 2021
  • Water quality prediction is essential for the proper management of water supply systems. Increased suspended sediment concentration (SSC) has various effects on water supply systems such as increased treatment cost and consequently, there have been various efforts to develop a model for predicting SSC. However, SSC is affected by both the natural and anthropogenic environment, making it challenging to predict SSC. Recently, advanced machine learning models have increasingly been used for water quality prediction. This study developed an ensemble machine learning model to predict SSC using the XGBoost (XGB) algorithm. The observed discharge (Q) and SSC in two fields monitoring stations were used to develop the model. The input variables were clustered in two groups with low and high ranges of Q using the k-means clustering algorithm. Then each group of data was separately used to optimize XGB (Model 1). The model performance was compared with that of the XGB model using the entire data (Model 2). The models were evaluated by mean squared error-ob servation standard deviation ratio (RSR) and root mean squared error. The RSR were 0.51 and 0.57 in the two monitoring stations for Model 2, respectively, while the model performance improved to RSR 0.46 and 0.55, respectively, for Model 1.

A Study of Machine Learning Model for Prediction of Swelling Waves Occurrence on East Sea (동해안 너울성 파도 예측을 위한 머신러닝 모델 연구)

  • Kang, Donghoon;Oh, Sejong
    • The Journal of Korean Institute of Information Technology
    • /
    • v.17 no.9
    • /
    • pp.11-17
    • /
    • 2019
  • In recent years, damage and loss of life and property have been occurred frequently due to swelling waves in the East Sea. Swelling waves are not easy to predict because they are caused by various factors. In this research, we build a model for predicting the swelling waves occurrence in the East Coast of Korea using machine learning technique. We collect historical data of unloading interruption in the Pohang Port, and collect air pressure, wind speed, direction, water temperature data of the offshore Pohang Port. We select important variables for prediction, and test various machine learning prediction algorithms. As a result, tide level, water temperature, and air pressure were selected, and Random Forest model produced best performance. We confirm that Random Forest model shows best performance and it produces 88.86% of accuracy

Machine Learning Methods for Trust-based Selection of Web Services

  • Hasnain, Muhammad;Ghani, Imran;Pasha, Muhammad F.;Jeong, Seung R.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.1
    • /
    • pp.38-59
    • /
    • 2022
  • Web services instances can be classified into two categories, namely trusted and untrusted from users. A web service with high throughput (TP) and low response time (RT) instance values is a trusted web service. Web services are not trustworthy due to the mismatch in the guaranteed instance values and the actual values achieved by users. To perform web services selection from users' attained TP and RT values, we need to verify the correct prediction of trusted and untrusted instances from invoked web services. This accurate prediction of web services instances is used to perform the selection of web services. We propose to construct fuzzy rules to label web services instances correctly. This paper presents web services selection using a well-known machine learning algorithm, namely REPTree, for the correct prediction of trusted and untrusted instances. Performance comparison of REPTree with five machine learning models is conducted on web services datasets. We have performed experiments on web services datasets using a ten k-fold cross-validation method. To evaluate the performance of the REPTree classifier, we used accuracy metrics (Sensitivity and Specificity). Experimental results showed that web service (WS1) gained top selection score with the (47.0588%) trusted instances, and web service (WS2) was selected the least with (25.00%) trusted instances. Evaluation results of the proposed web services selection approach were found as (asymptotic sig. = 0.019), demonstrating the relationship between final selection and recommended trust score of web services.

Runoff Prediction from Machine Learning Models Coupled with Empirical Mode Decomposition: A case Study of the Grand River Basin in Canada

  • Parisouj, Peiman;Jun, Changhyun;Nezhad, Somayeh Moghimi;Narimani, Roya
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.136-136
    • /
    • 2022
  • This study investigates the possibility of coupling empirical mode decomposition (EMD) for runoff prediction from machine learning (ML) models. Here, support vector regression (SVR) and convolutional neural network (CNN) were considered for ML algorithms. Precipitation (P), minimum temperature (Tmin), maximum temperature (Tmax) and their intrinsic mode functions (IMF) values were used for input variables at a monthly scale from Jan. 1973 to Dec. 2020 in the Grand river basin, Canada. The support vector machine-recursive feature elimination (SVM-RFE) technique was applied for finding the best combination of predictors among input variables. The results show that the proposed method outperformed the individual performance of SVR and CNN during the training and testing periods in the study area. According to the correlation coefficient (R), the EMD-SVR model outperformed the EMD-CNN model in both training and testing even though the CNN indicated a better performance than the SVR before using IMF values. The EMD-SVR model showed higher improvement in R value (38.7%) than that from the EMD-CNN model (7.1%). It should be noted that the coupled models of EMD-SVR and EMD-CNN represented much higher accuracy in runoff prediction with respect to the considered evaluation indicators, including root mean square error (RMSE) and R values.

  • PDF