• Title/Summary/Keyword: LSTM algorithm

Search Result 199, Processing Time 0.025 seconds

Prediction of Water Quality Factor for River Basin using RNN-LSTM Algorithm (RNN-LSTM 알고리즘을 이용한 하천의 수질인자 예측)

  • Lim, Hee Sung;An, Hyun Uk
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2020.06a
    • /
    • pp.219-219
    • /
    • 2020
  • 하천의 수질을 나타내는 환경지표 중 국가 TMS(Tele Monitoring system)의 수질측정망을 통해 관리되고 있는 지표로는 DO, BOD, COD, SS, TN, TP 등 여러 인자들이 있다. 이러한 수질인자는 하천의 자정작용에 있어 많은 영향을 나타내고 있다. 이를 활용한 경제적이고 합리적인 수질관리를 위해 하천의 자정작용을 활용하는 것이 중요하다. 생물학적 작용을 가장 효과적으로 활용하기 위해서는 수질오염 데이터에 기초한 수질예측을 채택하여 적절한 대책이 필요하다. 이를 위해서는 수질인자의 데이터를 측정하고 축적해 수질오염을 예측하는 것이 필수적인데, 실제적으로 수질인자의 일일 측정은 비용 관점에서 쉽게 접근할 수 없다. 본 연구에서는 시계열 학습으로 알려진 RNN-LSTM(Recurrent Neural Network-Long Term Memory) 알고리즘을 활용하여 기존에 측정된 수질인자의 데이터를 통해 시간당 및 일일 수질인자를 예측하려고 했다. 연구에 앞서, 기존에 시간단위로 측정된 수질인자 데이터의 이상 유무를 확인 후, 에러값은 제거하고 12시간 이하 데이터가 누락되었을 때는 선형 보간하여 데이터를 사용하고, 1일 데이터도 10일 이하 데이터가 누락되었을 때 선형 보간하여 데이터를 활용하여 수질인자를 예측하였다. 수질인자를 예측하기 위해 구글이 개발한 딥러닝 오픈소스 라이브러리인 텐서플로우를 활용하였고, 연구지역으로는 대한민국 부산에 위치한 온천천의 유역을 선정하였다. 수질인자 데이터 수집은 부산광역시에서 운영하는 보건환경정보 공개시스템의 자료를 활용하였다. 모델의 연구를 위해 하천의 수질인자, 기상자료 데이터를 입력자료로 활용하였다. 분석에서는 입력자료와, 반복횟수, 시계열의 길이 등을 조절해 수질 요인을 예측했고, 모델의 정확도도 분석하였다.

  • PDF

Comparison of artificial intelligence models reconstructing missing wind signals in deep-cutting gorges

  • Zhen Wang;Jinsong Zhu;Ziyue Lu;Zhitian Zhang
    • Wind and Structures
    • /
    • v.38 no.1
    • /
    • pp.75-91
    • /
    • 2024
  • Reliable wind signal reconstruction can be beneficial to the operational safety of long-span bridges. Non-Gaussian characteristics of wind signals make the reconstruction process challenging. In this paper, non-Gaussian wind signals are converted into a combined prediction of two kinds of features, actual wind speeds and wind angles of attack. First, two decomposition techniques, empirical mode decomposition (EMD) and variational mode decomposition (VMD), are introduced to decompose wind signals into intrinsic mode functions (IMFs) to reduce the randomness of wind signals. Their principles and applicability are also discussed. Then, four artificial intelligence (AI) algorithms are utilized for wind signal reconstruction by combining the particle swarm optimization (PSO) algorithm with back propagation neural network (BPNN), support vector regression (SVR), long short-term memory (LSTM) and bidirectional long short-term memory (Bi-LSTM), respectively. Measured wind signals from a bridge site in a deep-cutting gorge are taken as experimental subjects. The results showed that the reconstruction error of high-frequency components of EMD is too large. On the contrary, VMD fully extracts the multiscale rules of the signal, reduces the component complexity. The combination of VMD-PSO-Bi-LSTM is demonstrated to be the most effective among all hybrid models.

Performance Evaluation of LSTM-based PM2.5 Prediction Model for Learning Seasonal and Concentration-specific Data (계절별 데이터와 농도별 데이터의 학습에 대한 LSTM 기반의 PM2.5 예측 모델 성능 평가)

  • Yong-jin Jung;Chang-Heon Oh
    • Journal of Advanced Navigation Technology
    • /
    • v.28 no.1
    • /
    • pp.149-154
    • /
    • 2024
  • Research on particulate matter is advancing in real-time, and various methods are being studied to improve the accuracy of prediction models. Furthermore, studies that take into account various factors to understand the precise causes and impacts of particulate matter are actively being pursued. This paper trains an LSTM model using seasonal data and another LSTM model using concentration-based data. It compares and analyzes the PM2.5 prediction performance of the two models. To train the model, weather data and air pollutant data were collected. The collected data was then used to confirm the correlation with PM2.5. Based on the results of the correlation analysis, the data was structured for training and evaluation. The seasonal prediction model and the concentration-specific prediction model were designed using the LSTM algorithm. The performance of the prediction model was evaluated using accuracy, RMSE, and MAPE. As a result of the performance evaluation, the prediction model learned by concentration had an accuracy of 91.02% in the "bad" range of AQI. And overall, it performed better than the prediction model trained by season.

Drug-Drug Interaction Prediction Using Krill Herd Algorithm Based on Deep Learning Method

  • Al-Marghilani, Abdulsamad
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.6
    • /
    • pp.319-328
    • /
    • 2021
  • Parallel administration of numerous drugs increases Drug-Drug Interaction (DDI) because one drug might affect the activity of other drugs. DDI causes negative or positive impacts on therapeutic output. So there is a need to discover DDI to enhance the safety of consuming drugs. Though there are several DDI system exist to predict an interaction but nowadays it becomes impossible to maintain with a large number of biomedical texts which is getting increased rapidly. Mostly the existing DDI system address classification issues, and especially rely on handcrafted features, and some features which are based on particular domain tools. The objective of this paper to predict DDI in a way to avoid adverse effects caused by the consumed drugs, to predict similarities among the drug, Drug pair similarity calculation is performed. The best optimal weight is obtained with the support of KHA. LSTM function with weight obtained from KHA and makes bets prediction of DDI. Our methodology depends on (LSTM-KHA) for the detection of DDI. Similarities among the drugs are measured with the help of drug pair similarity calculation. KHA is used to find the best optimal weight which is used by LSTM to predict DDI. The experimental result was conducted on three kinds of dataset DS1 (CYP), DS2 (NCYP), and DS3 taken from the DrugBank database. To evaluate the performance of proposed work in terms of performance metrics like accuracy, recall, precision, F-measures, AUPR, AUC, and AUROC. Experimental results express that the proposed method outperforms other existing methods for predicting DDI. LSTMKHA produces reasonable performance metrics when compared to the existing DDI prediction model.

Application of Statistical and Machine Learning Techniques for Habitat Potential Mapping of Siberian Roe Deer in South Korea

  • Lee, Saro;Rezaie, Fatemeh
    • Proceedings of the National Institute of Ecology of the Republic of Korea
    • /
    • v.2 no.1
    • /
    • pp.1-14
    • /
    • 2021
  • The study has been carried out with an objective to prepare Siberian roe deer habitat potential maps in South Korea based on three geographic information system-based models including frequency ratio (FR) as a bivariate statistical approach as well as convolutional neural network (CNN) and long short-term memory (LSTM) as machine learning algorithms. According to field observations, 741 locations were reported as roe deer's habitat preferences. The dataset were divided with a proportion of 70:30 for constructing models and validation purposes. Through FR model, a total of 10 influential factors were opted for the modelling process, namely altitude, valley depth, slope height, topographic position index (TPI), topographic wetness index (TWI), normalized difference water index, drainage density, road density, radar intensity, and morphological feature. The results of variable importance analysis determined that TPI, TWI, altitude and valley depth have higher impact on predicting. Furthermore, the area under the receiver operating characteristic (ROC) curve was applied to assess the prediction accuracies of three models. The results showed that all the models almost have similar performances, but LSTM model had relatively higher prediction ability in comparison to FR and CNN models with the accuracy of 76% and 73% during the training and validation process. The obtained map of LSTM model was categorized into five classes of potentiality including very low, low, moderate, high and very high with proportions of 19.70%, 19.81%, 19.31%, 19.86%, and 21.31%, respectively. The resultant potential maps may be valuable to monitor and preserve the Siberian roe deer habitats.

Intelligent interface using hand gestures recognition based on artificial intelligence (인공지능 기반 손 체스처 인식 정보를 활용한 지능형 인터페이스)

  • Hangjun Cho;Junwoo Yoo;Eun Soo Kim;Young Jae Lee
    • Journal of Platform Technology
    • /
    • v.11 no.1
    • /
    • pp.38-51
    • /
    • 2023
  • We propose an intelligent interface algorithm using hand gesture recognition information based on artificial intelligence. This method is functionally an interface that recognizes various motions quickly and intelligently by using MediaPipe and artificial intelligence techniques such as KNN, LSTM, and CNN to track and recognize user hand gestures. To evaluate the performance of the proposed algorithm, it is applied to a self-made 2D top-view racing game and robot control. As a result of applying the algorithm, it was possible to control various movements of the virtual object in the game in detail and robustly. And the result of applying the algorithm to the robot control in the real world, it was possible to control movement, stop, left turn, and right turn. In addition, by controlling the main character of the game and the robot in the real world at the same time, the optimized motion was implemented as an intelligent interface for controlling the coexistence space of virtual and real world. The proposed algorithm enables sophisticated control according to natural and intuitive characteristics using the body and fine movement recognition of fingers, and has the advantage of being skilled in a short period of time, so it can be used as basic data for developing intelligent user interfaces.

  • PDF

Trading Algorithm Selection Using Time-Series Generative Adversarial Networks (TimeGAN을 활용한 트레이딩 알고리즘 선택)

  • Lee, Jae Yoon;Lee, Ju Hong;Choi, Bum Ghi;Song, Jae Won
    • Smart Media Journal
    • /
    • v.11 no.1
    • /
    • pp.38-45
    • /
    • 2022
  • A lot of research is being going until this day in order to obtain stable profit in the stock market. Trading algorithms are widely used, accounting for over 80% of the trading volume of the US stock market. Despite a lot of research, there is no trading algorithm that always shows good performance. In other words, there is no guarantee that an algorithm that performed well in the past will perform well in the future. The reason is that there are many factors that affect the stock price and there are uncertainties about the future. Therefore, in this paper, we propose a model using TimeGAN that predicts future returns well and selects algorithms that are expected to have high returns based on past records of the returns of algorithms. We use TimeGAN becasue it is probabilistic, whereas LSTM method predicts future time series data is deterministic. The advantage of TimeGAN probabilistic prediction is that it can reflect uncertainty about the future. As an experimental result, the method proposed in this paper achieves a high return with little volatility and shows superior results compared to many comparison algorithms.

Preliminary Study of Deep Learning-based Precipitation

  • Kim, Hee-Un;Bae, Tae-Suk
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.35 no.5
    • /
    • pp.423-430
    • /
    • 2017
  • Recently, data analysis research has been carried out using the deep learning technique in various fields such as image interpretation and/or classification. Various types of algorithms are being developed for many applications. In this paper, we propose a precipitation prediction algorithm based on deep learning with high accuracy in order to take care of the possible severe damage caused by climate change. Since the geographical and seasonal characteristics of Korea are clearly distinct, the meteorological factors have repetitive patterns in a time series. Since the LSTM (Long Short-Term Memory) is a powerful algorithm for consecutive data, it was used to predict precipitation in this study. For the numerical test, we calculated the PWV (Precipitable Water Vapor) based on the tropospheric delay of the GNSS (Global Navigation Satellite System) signals, and then applied the deep learning technique to the precipitation prediction. The GNSS data was processed by scientific software with the troposphere model of Saastamoinen and the Niell mapping function. The RMSE (Root Mean Squared Error) of the precipitation prediction based on LSTM performs better than that of ANN (Artificial Neural Network). By adding GNSS-based PWV as a feature, the over-fitting that is a latent problem of deep learning was prevented considerably as discussed in this study.

Sketch Recognition Using LSTM with Attention Mechanism and Minimum Cost Flow Algorithm

  • Nguyen-Xuan, Bac;Lee, Guee-Sang
    • International Journal of Contents
    • /
    • v.15 no.4
    • /
    • pp.8-15
    • /
    • 2019
  • This paper presents a solution of the 'Quick, Draw! Doodle Recognition Challenge' hosted by Google. Doodles are drawings comprised of concrete representational meaning or abstract lines creatively expressed by individuals. In this challenge, a doodle is presented as a sequence of sketches. From the view of at the sketch level, to learn the pattern of strokes representing a doodle, we propose a sequential model stacked with multiple convolution layers and Long Short-Term Memory (LSTM) cells following the attention mechanism [15]. From the view at the image level, we use multiple models pre-trained on ImageNet to recognize the doodle. Finally, an ensemble and a post-processing method using the minimum cost flow algorithm are introduced to combine multiple models in achieving better results. In this challenge, our solutions garnered 11th place among 1,316 teams. Our performance was 0.95037 MAP@3, only 0.4% lower than the winner. It demonstrates that our method is very competitive. The source code for this competition is published at: https://github.com/ngxbac/Kaggle-QuickDraw.

Design of Ballistic Calculation Model for Improving Accuracy of Naval Gun Firing based on Deep Learning

  • Oh, Moon-Tak
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.12
    • /
    • pp.11-18
    • /
    • 2021
  • This paper shows the applicability of deep learning algorithm in predicting target position and getting correction value of impact point in order to improve the accuracy of naval gun firing. Predicting target position, the proposed model using LSTM model and RN structure is expected to be more accurate than existing method using kalman filter. Getting correction value of impact point, the another proposed model suggests a reinforcement model that manages factors which is related in ballistic calculation as data set, and learns using the data set. The model is expected to reduce error of naval gun firing. Combining two models, a ballistic calculation model for improving accuracy of naval gun firing based on deep learning algorithm was designed.