• Title/Summary/Keyword: Long short time memory

Search Result 284, Processing Time 0.031 seconds

Stock Prediction Model based on Bidirectional LSTM Recurrent Neural Network (양방향 LSTM 순환신경망 기반 주가예측모델)

  • Joo, Il-Taeck;Choi, Seung-Ho
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.11 no.2
    • /
    • pp.204-208
    • /
    • 2018
  • In this paper, we proposed and evaluated the time series deep learning prediction model for learning fluctuation pattern of stock price. Recurrent neural networks, which can store previous information in the hidden layer, are suitable for the stock price prediction model, which is time series data. In order to maintain the long - term dependency by solving the gradient vanish problem in the recurrent neural network, we use LSTM with small memory inside the recurrent neural network. Furthermore, we proposed the stock price prediction model using bidirectional LSTM recurrent neural network in which the hidden layer is added in the reverse direction of the data flow for solving the limitation of the tendency of learning only based on the immediately preceding pattern of the recurrent neural network. In this experiment, we used the Tensorflow to learn the proposed stock price prediction model with stock price and trading volume input. In order to evaluate the performance of the stock price prediction, the mean square root error between the real stock price and the predicted stock price was obtained. As a result, the stock price prediction model using bidirectional LSTM recurrent neural network has improved prediction accuracy compared with unidirectional LSTM recurrent neural network.

Large-Scale Text Classification with Deep Neural Networks (깊은 신경망 기반 대용량 텍스트 데이터 분류 기술)

  • Jo, Hwiyeol;Kim, Jin-Hwa;Kim, Kyung-Min;Chang, Jeong-Ho;Eom, Jae-Hong;Zhang, Byoung-Tak
    • KIISE Transactions on Computing Practices
    • /
    • v.23 no.5
    • /
    • pp.322-327
    • /
    • 2017
  • The classification problem in the field of Natural Language Processing has been studied for a long time. Continuing forward with our previous research, which classifies large-scale text using Convolutional Neural Networks (CNN), we implemented Recurrent Neural Networks (RNN), Long-Short Term Memory (LSTM) and Gated Recurrent Units (GRU). The experiment's result revealed that the performance of classification algorithms was Multinomial Naïve Bayesian Classifier < Support Vector Machine (SVM) < LSTM < CNN < GRU, in order. The result can be interpreted as follows: First, the result of CNN was better than LSTM. Therefore, the text classification problem might be related more to feature extraction problem than to natural language understanding problems. Second, judging from the results the GRU showed better performance in feature extraction than LSTM. Finally, the result that the GRU was better than CNN implies that text classification algorithms should consider feature extraction and sequential information. We presented the results of fine-tuning in deep neural networks to provide some intuition regard natural language processing to future researchers.

Ultra-Low Powered CNT Synaptic Transistor Utilizing Double PI:PCBM Dielectric Layers (더블 PI:PCBM 유전체 층 기반의 초 저전력 CNT 시냅틱 트랜지스터)

  • Kim, Yonghun;Cho, Byungjin
    • Korean Journal of Materials Research
    • /
    • v.27 no.11
    • /
    • pp.590-596
    • /
    • 2017
  • We demonstrated a CNT synaptic transistor by integrating 6,6-phenyl-C61 butyric acid methyl ester(PCBM) molecules as charge storage molecules in a polyimide(PI) dielectric layer with carbon nanotubes(CNTs) for the transistor channel. Specifically, we fabricated and compared three different kinds of CNT-based synaptic transistors: a control device with $Al_2O_3/PI$, a single PCBM device with $Al_2O_3/PI:PCBM$(0.1 wt%), and a double PCBM device with $Al_2O_3/PI:PCBM$(0.1 wt%)/PI:PCBM(0.05 wt%). Statistically, essential device parameters such as Off and On currents, On/Off ratio, device yield, and long-term retention stability for the three kinds of transistor devices were extracted and compared. Notably, the double PCBM device exhibited the most excellent memory transistor behavior. Pulse response properties with postsynaptic dynamic current were also evaluated. Among all of the testing devices, double PCBM device consumed such low power for stand-by and its peak current ratio was so large that the postsynaptic current was also reliably and repeatedly generated. Postsynaptic hole currents through the CNT channel can be generated by electrons trapped in the PCBM molecules and last for a relatively short time(~ hundreds of msec). Under one certain testing configuration, the electrons trapped in the PCBM can also be preserved in a nonvolatile manner for a long-term period. Its integrated platform with extremely low stand-by power should pave a promising road toward next-generation neuromorphic systems, which would emulate the brain power of 20 W.

A Slot Allocated Blocking Anti-Collision Algorithm for RFID Tag Identification

  • Qing, Yang;Jiancheng, Li;Hongyi, Wang;Xianghua, Zeng;Liming, Zheng
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.6
    • /
    • pp.2160-2179
    • /
    • 2015
  • In many Radio Frequency Identification (RFID) applications, the reader recognizes the tags within its scope repeatedly. For these applications, some algorithms such as the adaptive query splitting algorithm (AQS) and the novel semi-blocking AQS (SBA) were proposed. In these algorithms, a staying tag retransmits its ID to the reader to be identified, even though the ID of the tag is stored in the reader's memory. When the length of tag ID is long, the reader consumes a long time to identify the staying tags. To overcome this deficiency, we propose a slot allocated blocking anti-collision algorithm (SABA). In SABA, the reader assigns a unique slot to each tag in its range by using a slot allocation mechanism. Based on the allocated slot, each staying tag only replies a short data to the reader in the identification process. As a result, the amount of data transmitted by the staying tags is reduced greatly and the identification rate of the reader is improved effectively. The identification rate and the data amount transmitted by tags of SABA are analyzed theoretically and verified by various simulations. The simulation and analysis results show that the performance of SABA is superior to the existing algorithms significantly.

A novel method for generation and prediction of crack propagation in gravity dams

  • Zhang, Kefan;Lu, Fangyun;Peng, Yong;Li, Xiangyu
    • Structural Engineering and Mechanics
    • /
    • v.81 no.6
    • /
    • pp.665-675
    • /
    • 2022
  • The safety problems of giant hydraulic structures such as dams caused by terrorist attacks, earthquakes, and wars often have an important impact on a country's economy and people's livelihood. For the national defense department, timely and effective assessment of damage to or impending damage to dams and other structures is an important issue related to the safety of people's lives and property. In the field of damage assessment and vulnerability analysis, it is usually necessary to give the damage assessment results within a few minutes to determine the physical damage (crack length, crater size, etc.) and functional damage (decreased power generation capacity, dam stability descent, etc.), so that other defense and security departments can take corresponding measures to control potential other hazards. Although traditional numerical calculation methods can accurately calculate the crack length and crater size under certain combat conditions, it usually takes a long time and is not suitable for rapid damage assessment. In order to solve similar problems, this article combines simulation calculation methods with machine learning technology interdisciplinary. First, the common concrete gravity dam shape was selected as the simulation calculation object, and XFEM (Extended Finite Element Method) was used to simulate and calculate 19 cracks with different initial positions. Then, an LSTM (Long-Short Term Memory) machine learning model was established. 15 crack paths were selected as the training set and others were set for test. At last, the LSTM model was trained by the training set, and the prediction results on the crack path were compared with the test set. The results show that this method can be used to predict the crack propagation path rapidly and accurately. In general, this article explores the application of machine learning related technologies in the field of mechanics. It has broad application prospects in the fields of damage assessment and vulnerability analysis.

Prediction in Dissolved Oxygen Concentration and Occurrence of Hypoxia Water Mass in Jinhae Bay Based on Machine Learning Model (기계학습 모형 기반 진해만 용존산소농도 및 빈산소수괴 발생 예측)

  • Park, Seongsik;Kim, Byeong Kuk;Kim, Kyunghoi
    • Journal of Korean Society of Coastal and Ocean Engineers
    • /
    • v.34 no.3
    • /
    • pp.47-57
    • /
    • 2022
  • We carried out studies on prediction in concentration of dissolved oxygen (DO) with LSTM model and prediction in occurrence of hypoxia water mass (HWM) with decision tree. As results of study on prediction in DO concentration, a large number of Hidden node caused high complexity of model and required enough Epoch. And it was high accuracy in long Sequence length as prediction time step increased. The results of prediction in occurrence of HWM showed that the accuracy of nonHWM case was 66.1% in 30 day prediction, it was higher than 37.5% of HWM case. The reason is that the decision tree might overestimate DO concentration.

Estimation of reaction forces at the seabed anchor of the submerged floating tunnel using structural pattern recognition

  • Seongi Min;Kiwon Jeong;Yunwoo Lee;Donghwi Jung;Seungjun Kim
    • Computers and Concrete
    • /
    • v.31 no.5
    • /
    • pp.405-417
    • /
    • 2023
  • The submerged floating tunnel (SFT) is tethered by mooring lines anchored to the seabed, therefore, the structural integrity of the anchor should be sensitively managed. Despite their importance, reaction forces cannot be simply measured by attaching sensors or load cells because of the structural and environmental characteristics of the submerged structure. Therefore, we propose an effective method for estimating the reaction forces at the seabed anchor of a submerged floating tunnel using a structural pattern model. First, a structural pattern model is established to use the correlation between tunnel motion and anchor reactions via a deep learning algorithm. Once the pattern model is established, it is directly used to estimate the reaction forces by inputting the tunnel motion data, which can be directly measured inside the tunnel. Because the sequential characteristics of responses in the time domain should be considered, the long short-term memory (LSTM) algorithm is mainly used to recognize structural behavioral patterns. Using hydrodynamics-based simulations, big data on the structural behavior of the SFT under various waves were generated, and the prepared datasets were used to validate the proposed method. The simulation-based validation results clearly show that the proposed method can precisely estimate time-series reactions using only acceleration data. In addition to real-time structural health monitoring, the proposed method can be useful for forensics when an unexpected accident or failure is related to the seabed anchors of the SFT.

Developing Cryptocurrency Trading Strategies with Time Series Forecasting Model (시계열 예측 모델을 활용한 암호화폐 투자 전략 개발)

  • Hyun-Sun Kim;Jae Joon Ahn
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.46 no.4
    • /
    • pp.152-159
    • /
    • 2023
  • This study endeavors to enrich investment prospects in cryptocurrency by establishing a rationale for investment decisions. The primary objective involves evaluating the predictability of four prominent cryptocurrencies - Bitcoin, Ethereum, Litecoin, and EOS - and scrutinizing the efficacy of trading strategies developed based on the prediction model. To identify the most effective prediction model for each cryptocurrency annually, we employed three methodologies - AutoRegressive Integrated Moving Average (ARIMA), Long Short-Term Memory (LSTM), and Prophet - representing traditional statistics and artificial intelligence. These methods were applied across diverse periods and time intervals. The result suggested that Prophet trained on the previous 28 days' price history at 15-minute intervals generally yielded the highest performance. The results were validated through a random selection of 100 days (20 target dates per year) spanning from January 1st, 2018, to December 31st, 2022. The trading strategies were formulated based on the optimal-performing prediction model, grounded in the simple principle of assigning greater weight to more predictable assets. When the forecasting model indicates an upward trend, it is recommended to acquire the cryptocurrency with the investment amount determined by its performance. Experimental results consistently demonstrated that the proposed trading strategy yields higher returns compared to an equal portfolio employing a buy-and-hold strategy. The cryptocurrency trading model introduced in this paper carries two significant implications. Firstly, it facilitates the evolution of cryptocurrencies from speculative assets to investment instruments. Secondly, it plays a crucial role in advancing deep learning-based investment strategies by providing sound evidence for portfolio allocation. This addresses the black box issue, a notable weakness in deep learning, offering increased transparency to the model.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

Digital Pre-Distortion Technique Using Repeated Usage of Feedback Samples (피드백 샘플 반복 활용을 이용한 다지털 전치 왜곡 방안)

  • Lee, Kwang-Pyo;Hong, Soon-Il;Jeong, Eui-Rim
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2015.05a
    • /
    • pp.673-676
    • /
    • 2015
  • Digital Pre-Distortion (DPD) is a linearization technique for nonlinear power amplifiers (PAs) by implementing inverse function of the PA at baseband digital stage. To obtain proper DPD parameters, a feedback path is required to convert the PA output to a baseband signal, and a memory is also needed to store the feedback signals. DPD parameters are usually found by an adaptive algorithm from the feedback samples. However, for the adaptive algorithm to converge to a reliable solution, long feedback samples are required, which increases convergence time and hardware complexity. In this paper, we propose a DPD technique that requires relatively short feedback samples. From the observation that the convergence time of the adaptive algorithm highly depends on the initial condition, this paper iteratively utilizes the feedback samples while keeping and using the converged DPD parameters at the former iteration as the initial condition at the current iteration. Computer simulation results show that the proposed method performs better than the conventional technique while the former requires much shorter feedback samples than the latter.

  • PDF