• Title/Summary/Keyword: recurrent neural network (RNN)

Search Result 228, Processing Time 0.021 seconds

Mid-Term Energy Demand Forecasting Using Conditional Restricted Boltzmann Machine (조건적 제한된 볼츠만머신을 이용한 중기 전력 수요 예측)

  • Kim, Soo-Hyun;Sun, Young-Ghyu;Lee, Dong-gu;Sim, Is-sac;Hwang, Yu-Min;Kim, Hyun-Soo;Kim, Hyung-suk;Kim, Jin-Young
    • Journal of IKEEE
    • /
    • v.23 no.1
    • /
    • pp.127-133
    • /
    • 2019
  • Electric power demand forecasting is one of the important research areas for future smart grid introduction. However, It is difficult to predict because it is affected by many external factors. Traditional methods of forecasting power demand have been limited in making accurate prediction because they use raw power data. In this paper, a probability-based CRBM is proposed to solve the problem of electric power demand prediction using raw power data. The stochastic model is suitable to capture the probabilistic characteristics of electric power data. In order to compare the mid-term power demand forecasting performance of the proposed model, we compared the performance with Recurrent Neural Network(RNN). Performance comparison using electric power data provided by the University of Massachusetts showed that the proposed algorithm results in better performance in mid-term energy demand forecasting.

Non-Intrusive Load Monitoring Method based on Long-Short Term Memory to classify Power Usage of Appliances (가전제품 전력 사용 분류를 위한 장단기 메모리 기반 비침입 부하 모니터링 기법)

  • Kyeong, Chanuk;Seon, Joonho;Sun, Young-Ghyu;Kim, Jin-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.4
    • /
    • pp.109-116
    • /
    • 2021
  • In this paper, we propose a non-intrusive load monitoring(NILM) system which can find the power of each home appliance from the aggregated total power as the activation in the trading market of the distributed resource and the increasing importance of energy management. We transform the amount of appliances' power into a power on-off state by preprocessing. We use LSTM as a model for predicting states based on these data. Accuracy is measured by comparing predicted states with real ones after postprocessing. In this paper, the accuracy is measured with the different number of electronic products, data postprocessing method, and Time step size. When the number of electronic products is 6, the data postprocessing method using the Round function is used, and Time step size is set to 6, the maximum accuracy can be obtained.

A Study on the Index Estimation of Missing Real Estate Transaction Cases Using Machine Learning (머신러닝을 활용한 결측 부동산 매매 지수의 추정에 대한 연구)

  • Kim, Kyung-Min;Kim, Kyuseok;Nam, Daisik
    • Journal of the Economic Geographical Society of Korea
    • /
    • v.25 no.1
    • /
    • pp.171-181
    • /
    • 2022
  • The real estate price index plays key roles as quantitative data in real estate market analysis. International organizations including OECD publish the real estate price indexes by country, and the Korea Real Estate Board announces metropolitan-level and municipal-level indexes. However, when the index is set on the smaller spatial unit level than metropolitan and municipal-level, problems occur: missing values. As the spatial scope is narrowed down, there are cases where there are few or no transactions depending on the unit period, which lead index calculation difficult or even impossible. This study suggests a supervised learning-based machine learning model to compensate for missing values that may occur due to no transaction in a specific range and period. The models proposed in our research verify the accuracy of predicting the existing values and missing values.

Forecasting Baltic Dry Index by Implementing Time-Series Decomposition and Data Augmentation Techniques (시계열 분해 및 데이터 증강 기법 활용 건화물운임지수 예측)

  • Han, Min Soo;Yu, Song Jin
    • Journal of Korean Society for Quality Management
    • /
    • v.50 no.4
    • /
    • pp.701-716
    • /
    • 2022
  • Purpose: This study aims to predict the dry cargo transportation market economy. The subject of this study is the BDI (Baltic Dry Index) time-series, an index representing the dry cargo transport market. Methods: In order to increase the accuracy of the BDI time-series, we have pre-processed the original time-series via time-series decomposition and data augmentation techniques and have used them for ANN learning. The ANN algorithms used are Multi-Layer Perceptron (MLP), Recurrent Neural Network (RNN), and Long Short-Term Memory (LSTM) to compare and analyze the case of learning and predicting by applying time-series decomposition and data augmentation techniques. The forecast period aims to make short-term predictions at the time of t+1. The period to be studied is from '22. 01. 07 to '22. 08. 26. Results: Only for the case of the MAPE (Mean Absolute Percentage Error) indicator, all ANN models used in the research has resulted in higher accuracy (1.422% on average) in multivariate prediction. Although it is not a remarkable improvement in prediction accuracy compared to uni-variate prediction results, it can be said that the improvement in ANN prediction performance has been achieved by utilizing time-series decomposition and data augmentation techniques that were significant and targeted throughout this study. Conclusion: Nevertheless, due to the nature of ANN, additional performance improvements can be expected according to the adjustment of the hyper-parameter. Therefore, it is necessary to try various applications of multiple learning algorithms and ANN optimization techniques. Such an approach would help solve problems with a small number of available data, such as the rapidly changing business environment or the current shipping market.

Semantic analysis via application of deep learning using Naver movie review data (네이버 영화 리뷰 데이터를 이용한 의미 분석(semantic analysis))

  • Kim, Sojin;Song, Jongwoo
    • The Korean Journal of Applied Statistics
    • /
    • v.35 no.1
    • /
    • pp.19-33
    • /
    • 2022
  • With the explosive growth of social media, its abundant text-based data generated by web users has become an important source for data analysis. For example, we often witness online movie reviews from the 'Naver Movie' affecting the general public to decide whether they should watch the movie or not. This study has conducted analysis on the Naver Movie's text-based review data to predict the actual ratings. After examining the distribution of movie ratings, we performed semantics analysis using Korean Natural Language Processing. This research sought to find the best review rating prediction model by comparing machine learning and deep learning models. We also compared various regression and classification models in 2-class and multi-class cases. Lastly we explained the causes of review misclassification related to movie review data characteristics.

Very Short- and Long-Term Prediction Method for Solar Power (초 장단기 통합 태양광 발전량 예측 기법)

  • Mun Seop Yun;Se Ryung Lim;Han Seung Jang
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.6
    • /
    • pp.1143-1150
    • /
    • 2023
  • The global climate crisis and the implementation of low-carbon policies have led to a growing interest in renewable energy and a growing number of related industries. Among them, solar power is attracting attention as a representative eco-friendly energy that does not deplete and does not emit pollutants or greenhouse gases. As a result, the supplement of solar power facility is increasing all over the world. However, solar power is easily affected by the environment such as geography and weather, so accurate solar power forecast is important for stable operation and efficient management. However, it is very hard to predict the exact amount of solar power using statistical methods. In addition, the conventional prediction methods have focused on only short- or long-term prediction, which causes to take long time to obtain various prediction models with different prediction horizons. Therefore, this study utilizes a many-to-many structure of a recurrent neural network (RNN) to integrate short-term and long-term predictions of solar power generation. We compare various RNN-based very short- and long-term prediction methods for solar power in terms of MSE and R2 values.

A Study on the Settlement Prediction of Soft Ground Embankment Using Artificial Neural Network (인공신경망을 이용한 연약지반성토의 침하예측 연구)

  • Kim, Dong-Sik;Chae, Young-Su;Kim, Young-Su;Kim, Hyun-Dong
    • Journal of the Korean Geotechnical Society
    • /
    • v.23 no.7
    • /
    • pp.17-25
    • /
    • 2007
  • Various geotechnical problems due to insufficient bearing capacity or excessive settlement are likely to occur when constructing roads or large complexes on soft ground. Accurate predictions of the magnitude of settlement and the consolidation time provide numerous options of ground improvement methods and, thus, enable to save time and expense of the whole project. Asaoka's method is probably the most frequently used one for settlement prediction and the empirical formulae such as Hyperbolic method and Hoshino's method are also often used. To find an elaborate method of predicting the embankment settlement, two recurrent type neural network models, such as Jordan model and Elman-Jordan model, are adopted. The data sets of settlement measured at several domestic sites are analyzed to obtain the most suitable model structures. It was shown from the comparison between predicted and measured settlements that Jordan model provides better predictions than Elman-Jordan model does and that the predictions using CPT results are more accurate than those using SPT results. It is believed that RNN using cone penetration test results can be a highly efficient tool in predicting settlements if enough field data can be obtained.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.