• Title/Summary/Keyword: Long short term memory (LSTM)

Search Result 522, Processing Time 0.028 seconds

Forecasting Container Throughput with Long Short Term Memory (LSTM을 활용한 컨테이너 물동량 예측)

  • Lim, Sangseop
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2020.07a
    • /
    • pp.617-618
    • /
    • 2020
  • 우리나라의 지리적인 여건상 대륙과 연결되지 않기 때문에 해상운송에 절대적으로 의존하고 있다. 해상운송에 있어 항만시설의 확보가 필요하며 대외무역의존도가 높은 우리나라의 경우 더욱 중요한 역할을 한다. 항만시설은 장기적인 항만수요예측을 통해 대규모 인프라투자를 결정하며 단기적인 예측은 항만운영의 효율성을 개선하고 항만의 경쟁력을 제고하는데 기여하므로 예측의 정확성을 높이기 위해 많은 노력이 필요하다. 본 논문에서는 딥러닝 모델 중에 하나인 LSTM(Long Short Term Memory)을 적용하여 우리나라 주요항만의 컨테이너 물동량 단기예측을 수행하여 선행연구들에서 주류를 이뤘던 ARIMA류의 시계열모델과 비교하여 예측성능을 평가할 것이다. 본 논문은 학문적으로 항만수요예측에 관한 새로운 예측모델을 제시하였다는 측면에서 의미가 있으며 실무적으로 항만수요예측에 대한 정확성을 개선하여 항만투자의사결정에 과학적인 근거로서 활용이 가능할 것으로 기대된다.

  • PDF

Text Classification Method Using Deep Learning Model Fusion and Its Application

  • Shin, Seong-Yoon;Cho, Gwang-Hyun;Cho, Seung-Pyo;Lee, Hyun-Chang
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.409-410
    • /
    • 2022
  • This paper proposes a fusion model based on Long-Short Term Memory networks (LSTM) and CNN deep learning methods, and applied to multi-category news datasets, and achieved good results. Experiments show that the fusion model based on deep learning has greatly improved the precision and accuracy of text sentiment classification. This method will become an important way to optimize the model and improve the performance of the model.

  • PDF

Remaining Useful Life Prediction for Litium-Ion Batteries Using EMD-CNN-LSTM Hybrid Method (EMD-CNN-LSTM을 이용한 하이브리드 방식의 리튬 이온 배터리 잔여 수명 예측)

  • Lim, Je-Yeong;Kim, Dong-Hwan;Noh, Tae-Won;Lee, Byoung-Kuk
    • The Transactions of the Korean Institute of Power Electronics
    • /
    • v.27 no.1
    • /
    • pp.48-55
    • /
    • 2022
  • This paper proposes a battery remaining useful life (RUL) prediction method using a deep learning-based EMD-CNN-LSTM hybrid method. The proposed method pre-processes capacity data by applying empirical mode decomposition (EMD) and predicts the remaining useful life using CNN-LSTM. CNN-LSTM is a hybrid method that combines convolution neural network (CNN), which analyzes spatial features, and long short term memory (LSTM), which is a deep learning technique that processes time series data analysis. The performance of the proposed remaining useful life prediction method is verified using the battery aging experiment data provided by the NASA Ames Prognostics Center of Excellence and shows higher accuracy than does the conventional method.

Prediction of Sea Surface Temperature and Detection of Ocean Heat Wave in the South Sea of Korea Using Time-series Deep-learning Approaches (시계열 기계학습을 이용한 한반도 남해 해수면 온도 예측 및 고수온 탐지)

  • Jung, Sihun;Kim, Young Jun;Park, Sumin;Im, Jungho
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1077-1093
    • /
    • 2020
  • Sea Surface Temperature (SST) is an important environmental indicator that affects climate coupling systems around the world. In particular, coastal regions suffer from abnormal SST resulting in huge socio-economic damage. This study used Long Short Term Memory (LSTM) and Convolutional Long Short Term Memory (ConvLSTM) to predict SST up to 7 days in the south sea region in South Korea. The results showed that the ConvLSTM model outperformed the LSTM model, resulting in a root mean square error (RMSE) of 0.33℃ and a mean difference of -0.0098℃. Seasonal comparison also showed the superiority of ConvLSTM to LSTM for all seasons. However, in summer, the prediction accuracy for both models with all lead times dramatically decreased, resulting in RMSEs of 0.48℃ and 0.27℃ for LSTM and ConvLSTM, respectively. This study also examined the prediction of abnormally high SST based on three ocean heatwave categories (i.e., warning, caution, and attention) with the lead time from one to seven days for an ocean heatwave case in summer 2017. ConvLSTM was able to successfully predict ocean heatwave five days in advance.

Prediction of Baltic Dry Index by Applications of Long Short-Term Memory (Long Short-Term Memory를 활용한 건화물운임지수 예측)

  • HAN, Minsoo;YU, Song-Jin
    • Journal of Korean Society for Quality Management
    • /
    • v.47 no.3
    • /
    • pp.497-508
    • /
    • 2019
  • Purpose: The purpose of this study is to overcome limitations of conventional studies that to predict Baltic Dry Index (BDI). The study proposed applications of Artificial Neural Network (ANN) named Long Short-Term Memory (LSTM) to predict BDI. Methods: The BDI time-series prediction was carried out through eight variables related to the dry bulk market. The prediction was conducted in two steps. First, identifying the goodness of fitness for the BDI time-series of specific ANN models and determining the network structures to be used in the next step. While using ANN's generalization capability, the structures determined in the previous steps were used in the empirical prediction step, and the sliding-window method was applied to make a daily (one-day ahead) prediction. Results: At the empirical prediction step, it was possible to predict variable y(BDI time series) at point of time t by 8 variables (related to the dry bulk market) of x at point of time (t-1). LSTM, known to be good at learning over a long period of time, showed the best performance with higher predictive accuracy compared to Multi-Layer Perceptron (MLP) and Recurrent Neural Network (RNN). Conclusion: Applying this study to real business would require long-term predictions by applying more detailed forecasting techniques. I hope that the research can provide a point of reference in the dry bulk market, and furthermore in the decision-making and investment in the future of the shipping business as a whole.

Innovative Solutions for Design and Fabrication of Deep Learning Based Soft Sensor

  • Khdhir, Radhia;Belghith, Aymen
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.2
    • /
    • pp.131-138
    • /
    • 2022
  • Soft sensors are used to anticipate complicated model parameters using data from classifiers that are comparatively easy to gather. The goal of this study is to use artificial intelligence techniques to design and build soft sensors. The combination of a Long Short-Term Memory (LSTM) network and Grey Wolf Optimization (GWO) is used to create a unique soft sensor. LSTM is developed to tackle linear model with strong nonlinearity and unpredictability of manufacturing applications in the learning approach. GWO is used to accomplish input optimization technique for LSTM in order to reduce the model's inappropriate complication. The newly designed soft sensor originally brought LSTM's superior dynamic modeling with GWO's exact variable selection. The performance of our proposal is demonstrated using simulations on real-world datasets.

DG-based SPO tuple recognition using self-attention M-Bi-LSTM

  • Jung, Joon-young
    • ETRI Journal
    • /
    • v.44 no.3
    • /
    • pp.438-449
    • /
    • 2022
  • This study proposes a dependency grammar-based self-attention multilayered bidirectional long short-term memory (DG-M-Bi-LSTM) model for subject-predicate-object (SPO) tuple recognition from natural language (NL) sentences. To add recent knowledge to the knowledge base autonomously, it is essential to extract knowledge from numerous NL data. Therefore, this study proposes a high-accuracy SPO tuple recognition model that requires a small amount of learning data to extract knowledge from NL sentences. The accuracy of SPO tuple recognition using DG-M-Bi-LSTM is compared with that using NL-based self-attention multilayered bidirectional LSTM, DG-based bidirectional encoder representations from transformers (BERT), and NL-based BERT to evaluate its effectiveness. The DG-M-Bi-LSTM model achieves the best results in terms of recognition accuracy for extracting SPO tuples from NL sentences even if it has fewer deep neural network (DNN) parameters than BERT. In particular, its accuracy is better than that of BERT when the learning data are limited. Additionally, its pretrained DNN parameters can be applied to other domains because it learns the structural relations in NL sentences.

Effect of CAPPI Structure on the Perfomance of Radar Quantitative Precipitation Estimation using Long Short-Term Memory Networks

  • Dinh, Thi-Linh;Bae, Deg-Hyo
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.133-133
    • /
    • 2021
  • The performance of radar Quantitative Precipitation Estimation (QPE) using Long Short-Term Memory (LSTM) networks in hydrological applications depends on either the quality of data or the three-dimensional CAPPI structure from the weather radar. While radar data quality is controlled and enhanced by the more and more modern radar systems, the effect of CAPPI structure still has not yet fully investigated. In this study, three typical and important types of CAPPI structure including inverse-pyramid, cubic of grids 3x3, cubic of grids 4x4 are investigated to evaluate the effect of CAPPI structures on the performance of radar QPE using LSTM networks. The investigation results figure out that the cubic of grids 4x4 of CAPPI structure shows the best performance in rainfall estimation using the LSTM networks approach. This study give us the precious experiences in radar QPE works applying LSTM networks approach in particular and deep-learning approach in general.

  • PDF

The roles of differencing and dimension reduction in machine learning forecasting of employment level using the FRED big data

  • Choi, Ji-Eun;Shin, Dong Wan
    • Communications for Statistical Applications and Methods
    • /
    • v.26 no.5
    • /
    • pp.497-506
    • /
    • 2019
  • Forecasting the U.S. employment level is made using machine learning methods of the artificial neural network: deep neural network, long short term memory (LSTM), gated recurrent unit (GRU). We consider the big data of the federal reserve economic data among which 105 important macroeconomic variables chosen by McCracken and Ng (Journal of Business and Economic Statistics, 34, 574-589, 2016) are considered as predictors. We investigate the influence of the two statistical issues of the dimension reduction and time series differencing on the machine learning forecast. An out-of-sample forecast comparison shows that (LSTM, GRU) with differencing performs better than the autoregressive model and the dimension reduction improves long-term forecasts and some short-term forecasts.

Cross-Domain Text Sentiment Classification Method Based on the CNN-BiLSTM-TE Model

  • Zeng, Yuyang;Zhang, Ruirui;Yang, Liang;Song, Sujuan
    • Journal of Information Processing Systems
    • /
    • v.17 no.4
    • /
    • pp.818-833
    • /
    • 2021
  • To address the problems of low precision rate, insufficient feature extraction, and poor contextual ability in existing text sentiment analysis methods, a mixed model account of a CNN-BiLSTM-TE (convolutional neural network, bidirectional long short-term memory, and topic extraction) model was proposed. First, Chinese text data was converted into vectors through the method of transfer learning by Word2Vec. Second, local features were extracted by the CNN model. Then, contextual information was extracted by the BiLSTM neural network and the emotional tendency was obtained using softmax. Finally, topics were extracted by the term frequency-inverse document frequency and K-means. Compared with the CNN, BiLSTM, and gate recurrent unit (GRU) models, the CNN-BiLSTM-TE model's F1-score was higher than other models by 0.0147, 0.006, and 0.0052, respectively. Then compared with CNN-LSTM, LSTM-CNN, and BiLSTM-CNN models, the F1-score was higher by 0.0071, 0.0038, and 0.0049, respectively. Experimental results showed that the CNN-BiLSTM-TE model can effectively improve various indicators in application. Lastly, performed scalability verification through a takeaway dataset, which has great value in practical applications.