• Title/Summary/Keyword: optimize hyper-parameters

Search Result 9, Processing Time 0.021 seconds

A Study on Abnormal Data Processing Process of LSTM AE - With applying Data based Intelligent Factory

  • Youn-A Min
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.2
    • /
    • pp.240-247
    • /
    • 2023
  • In this paper, effective data management in industrial sites such as intelligent factories using time series data was studied. For effective management of time series data, variables considering the significance of the data were used, and hyper parameters calculated through LSTM AE were applied. We propose an optimized modeling considering the importance of each data section, and through this, outlier data of time series data can be efficiently processed. In the case of applying data significance and applying hyper parameters to which the research in this paper was applied, it was confirmed that the error rate was measured at 5.4%/4.8%/3.3%, and the significance of each data section and the significance of applying hyper parameters to optimize modeling were confirmed.

Optimization of Ground Contact Model of Ankleless Lower Exoskeleton Robot for Gait Simulation (보행 모의 실험을 위한 발목 없는 하지 외골격 로봇의 지면 접촉 모델 최적화)

  • Gimyeong Choi;Sanghyung Kim;Changhyun Cho
    • The Journal of Korea Robotics Society
    • /
    • v.18 no.4
    • /
    • pp.481-486
    • /
    • 2023
  • The purpose of this study is to optimize parameters of a contact model to obtain similar ground contact force of human walking. Dynamic walking simulation considering ground contact is performed to determine load specifications when developing walking assist robots. Large contact forces that are not observed in actual experimental data occur during the simulation at the initial contact (e.g., heel contact). The large contact force generates unrealistic large joint torques. A lower exoskeleton robot with no ankles is developed with the Matlab simscape and the nonlinear hyper volumetric contact model is applied. Parameters of the nonlinear hyper volumetric model were optimized using actual walking contact force data. As a result of optimization, it was possible to obtain a contact force pattern similar to actual walking by removing the large contact force generated during initial contact.

Feasibility Study of Google's Teachable Machine in Diagnosis of Tooth-Marked Tongue

  • Jeong, Hyunja
    • Journal of dental hygiene science
    • /
    • v.20 no.4
    • /
    • pp.206-212
    • /
    • 2020
  • Background: A Teachable Machine is a kind of machine learning web-based tool for general persons. In this paper, the feasibility of Google's Teachable Machine (ver. 2.0) was studied in the diagnosis of the tooth-marked tongue. Methods: For machine learning of tooth-marked tongue diagnosis, a total of 1,250 tongue images were used on Kaggle's web site. Ninety percent of the images were used for the training data set, and the remaining 10% were used for the test data set. Using Google's Teachable Machine (ver. 2.0), machine learning was performed using separated images. To optimize the machine learning parameters, I measured the diagnosis accuracies according to the value of epoch, batch size, and learning rate. After hyper-parameter tuning, the ROC (receiver operating characteristic) analysis method determined the sensitivity (true positive rate, TPR) and specificity (false positive rate, FPR) of the machine learning model to diagnose the tooth-marked tongue. Results: To evaluate the usefulness of the Teachable Machine in clinical application, I used 634 tooth-marked tongue images and 491 no-marked tongue images for machine learning. When the epoch, batch size, and learning rate as hyper-parameters were 75, 0.0001, and 128, respectively, the accuracy of the tooth-marked tongue's diagnosis was best. The accuracies for the tooth-marked tongue and the no-marked tongue were 92.1% and 72.6%, respectively. And, the sensitivity (TPR) and specificity (FPR) were 0.92 and 0.28, respectively. Conclusion: These results are more accurate than Li's experimental results calculated with convolution neural network. Google's Teachable Machines show good performance by hyper-parameters tuning in the diagnosis of the tooth-marked tongue. We confirmed that the tool is useful for several clinical applications.

Power consumption prediction model based on artificial neural networks for seawater source heat pump system in recirculating aquaculture system fish farm (순환여과식 양식장 해수 열원 히트펌프 시스템의 전력 소비량 예측을 위한 인공 신경망 모델)

  • Hyeon-Seok JEONG;Jong-Hyeok RYU;Seok-Kwon JEONG
    • Journal of the Korean Society of Fisheries and Ocean Technology
    • /
    • v.60 no.1
    • /
    • pp.87-99
    • /
    • 2024
  • This study deals with the application of an artificial neural network (ANN) model to predict power consumption for utilizing seawater source heat pumps of recirculating aquaculture system. An integrated dynamic simulation model was constructed using the TRNSYS program to obtain input and output data for the ANN model to predict the power consumption of the recirculating aquaculture system with a heat pump system. Data obtained from the TRNSYS program were analyzed using linear regression, and converted into optimal data necessary for the ANN model through normalization. To optimize the ANN-based power consumption prediction model, the hyper parameters of ANN were determined using the Bayesian optimization. ANN simulation results showed that ANN models with optimized hyper parameters exhibited acceptably high predictive accuracy conforming to ASHRAE standards.

Optimize rainfall prediction utilize multivariate time series, seasonal adjustment and Stacked Long short term memory

  • Nguyen, Thi Huong;Kwon, Yoon Jeong;Yoo, Je-Ho;Kwon, Hyun-Han
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.373-373
    • /
    • 2021
  • Rainfall forecasting is an important issue that is applied in many areas, such as agriculture, flood warning, and water resources management. In this context, this study proposed a statistical and machine learning-based forecasting model for monthly rainfall. The Bayesian Gaussian process was chosen to optimize the hyperparameters of the Stacked Long Short-term memory (SLSTM) model. The proposed SLSTM model was applied for predicting monthly precipitation of Seoul station, South Korea. Data were retrieved from the Korea Meteorological Administration (KMA) in the period between 1960 and 2019. Four schemes were examined in this study: (i) prediction with only rainfall; (ii) with deseasonalized rainfall; (iii) with rainfall and minimum temperature; (iv) with deseasonalized rainfall and minimum temperature. The error of predicted rainfall based on the root mean squared error (RMSE), 16-17 mm, is relatively small compared with the average monthly rainfall at Seoul station is 117mm. The results showed scheme (iv) gives the best prediction result. Therefore, this approach is more straightforward than the hydrological and hydraulic models, which request much more input data. The result indicated that a deep learning network could be applied successfully in the hydrology field. Overall, the proposed method is promising, given a good solution for rainfall prediction.

  • PDF

Tension Estimation of Tire using Neural Networks and DOE (신경회로망과 실험계획법을 이용한 타이어의 장력 추정)

  • Lee, Dong-Woo;Cho, Seok-Swoo
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.28 no.7
    • /
    • pp.814-820
    • /
    • 2011
  • It takes long time in numerical simulation because structural design for tire requires the nonlinear material property. Neural networks has been widely studied to engineering design to reduce numerical computation time. The numbers of hidden layer, hidden layer neuron and training data have been considered as the structural design variables of neural networks. In application of neural networks to optimize design, there are a few studies about arrangement method of input layer neurons. To investigate the effect of input layer neuron arrangement on neural networks, the variables of tire contour design and tension in bead area were assigned to inputs and output for neural networks respectively. Design variables arrangement in input layer were determined by main effect analysis. The number of hidden layer, the number of hidden layer neuron and the number of training data and so on have been considered as the structural design variables of neural networks. In application to optimization design problem of neural networks, there are few studies about arrangement method of input layer neurons. To investigate the effect of arrangement of input neurons on neural network learning tire contour design parameters and tension in bead area were assigned to neural input and output respectively. Design variables arrangement in input layer was determined by main effect analysis.

Reward Design of Reinforcement Learning for Development of Smart Control Algorithm (스마트 제어알고리즘 개발을 위한 강화학습 리워드 설계)

  • Kim, Hyun-Su;Yoon, Ki-Yong
    • Journal of Korean Association for Spatial Structures
    • /
    • v.22 no.2
    • /
    • pp.39-46
    • /
    • 2022
  • Recently, machine learning is widely used to solve optimization problems in various engineering fields. In this study, machine learning is applied to development of a control algorithm for a smart control device for reduction of seismic responses. For this purpose, Deep Q-network (DQN) out of reinforcement learning algorithms was employed to develop control algorithm. A single degree of freedom (SDOF) structure with a smart tuned mass damper (TMD) was used as an example structure. A smart TMD system was composed of MR (magnetorheological) damper instead of passive damper. Reward design of reinforcement learning mainly affects the control performance of the smart TMD. Various hyper-parameters were investigated to optimize the control performance of DQN-based control algorithm. Usually, decrease of the time step for numerical simulation is desirable to increase the accuracy of simulation results. However, the numerical simulation results presented that decrease of the time step for reward calculation might decrease the control performance of DQN-based control algorithm. Therefore, a proper time step for reward calculation should be selected in a DQN training process.

Fair Performance Evaluation Method for Stock Trend Prediction Models (주가 경향 예측 모델의 공정한 성능 평가 방법)

  • Lim, Chungsoo
    • The Journal of the Korea Contents Association
    • /
    • v.20 no.10
    • /
    • pp.702-714
    • /
    • 2020
  • Stock investment is a personal investment technique that has gathered tremendous interest since the reduction in interest rates and tax exemption. However, it is risky especially for those who do not have expert knowledge on stock volatility. Therefore, it is well understood that accurate stock trend prediction can greatly help stock investment, giving birth to a volume of research work in the field. In order to compare different research works and to optimize hyper-parameters for prediction models, it is required to have an evaluation standard that can accurately assess performances of prediction models. However, little research has been done in the area, and conventionally used methods have been employed repeatedly without being rigorously validated. For this reason, we first analyze performance evaluation of stock trend prediction with respect to performance metrics and data composition, and propose a fair evaluation method based on prediction disparity ratio.

Performance Evaluation of Machine Learning and Deep Learning Algorithms in Crop Classification: Impact of Hyper-parameters and Training Sample Size (작물분류에서 기계학습 및 딥러닝 알고리즘의 분류 성능 평가: 하이퍼파라미터와 훈련자료 크기의 영향 분석)

  • Kim, Yeseul;Kwak, Geun-Ho;Lee, Kyung-Do;Na, Sang-Il;Park, Chan-Won;Park, No-Wook
    • Korean Journal of Remote Sensing
    • /
    • v.34 no.5
    • /
    • pp.811-827
    • /
    • 2018
  • The purpose of this study is to compare machine learning algorithm and deep learning algorithm in crop classification using multi-temporal remote sensing data. For this, impacts of machine learning and deep learning algorithms on (a) hyper-parameter and (2) training sample size were compared and analyzed for Haenam-gun, Korea and Illinois State, USA. In the comparison experiment, support vector machine (SVM) was applied as machine learning algorithm and convolutional neural network (CNN) was applied as deep learning algorithm. In particular, 2D-CNN considering 2-dimensional spatial information and 3D-CNN with extended time dimension from 2D-CNN were applied as CNN. As a result of the experiment, it was found that the hyper-parameter values of CNN, considering various hyper-parameter, defined in the two study areas were similar compared with SVM. Based on this result, although it takes much time to optimize the model in CNN, it is considered that it is possible to apply transfer learning that can extend optimized CNN model to other regions. Then, in the experiment results with various training sample size, the impact of that on CNN was larger than SVM. In particular, this impact was exaggerated in Illinois State with heterogeneous spatial patterns. In addition, the lowest classification performance of 3D-CNN was presented in Illinois State, which is considered to be due to over-fitting as complexity of the model. That is, the classification performance was relatively degraded due to heterogeneous patterns and noise effect of input data, although the training accuracy of 3D-CNN model was high. This result simply that a proper classification algorithms should be selected considering spatial characteristics of study areas. Also, a large amount of training samples is necessary to guarantee higher classification performance in CNN, particularly in 3D-CNN.