• Title/Summary/Keyword: long term neural network

Search Result 395, Processing Time 0.021 seconds

Comparison of Fault Diagnosis Accuracy Between XGBoost and Conv1D Using Long-Term Operation Data of Ship Fuel Supply Instruments (선박 연료 공급 기기류의 장시간 운전 데이터의 고장 진단에 있어서 XGBoost 및 Conv1D의 예측 정확성 비교)

  • Hyung-Jin Kim;Kwang-Sik Kim;Se-Yun Hwang;Jang-Hyun Lee
    • Proceedings of the Korean Institute of Navigation and Port Research Conference
    • /
    • 2022.06a
    • /
    • pp.110-110
    • /
    • 2022
  • 본 연구는 자율운항 선박의 원격 고장 진단 기법 개발의 일부로 수행되었다. 특히, 엔진 연료 계통 장비로부터 계측된 시계열 데이터로부터 상태 진단을 위한 알고리즘 구현 결과를 제시하였다. 엔진 연료 펌프와 청정기를 가진 육상 실험 장비로부터 진동 시계열 데이터 계측하였으며, 이상 감지, 고장 분류 및 고장 예측이 가능한 심층 학습(Deep Learning) 및 기계 학습(Machine Learning) 알고리즘을 구현하였다. 육상 실험 장비에 고장 유형 별로 인위적인 고장을 발생시켜 특징적인 진동 신호를 계측하여, 인공 지능 학습에 이용하였다. 계측된 신호 데이터는 선행 발생한 사건의 신호가 후행 사건에 영향을 미치는 특성을 가지고 있으므로, 시계열에 내포된 고장 상태는 시간 간의 선후 종속성을 반영할 수 있는 학습 알고리즘을 제시하였다. 고장 사건의 시간 종속성을 반영할 수 있도록 순환(Recurrent) 계열의 RNN(Recurrent Neural Networks), LSTM(Long Short-Term Memory models)의 모델과 합성곱 연산 (Convolution Neural Network)을 기반으로 하는 Conv1D 모델을 적용하여 예측 정확성을 비교하였다. 특히, 합성곱 계열의 RNN LSTM 모델이 고차원의 순차적 자연어 언어 처리에 장점을 보이는 모델임을 착안하여, 신호의 시간 종속성을 학습에 반영할 수 있는 합성곱 계열의 Conv1 알고리즘을 고장 예측에 사용하였다. 또한 기계 학습 모델의 효율성을 감안하여 XGBoost를 추가로 적용하여 고장 예측을 시도하였다. 최종적으로 연료 펌프와 청정기의 진동 신호로부터 Conv1D 모델과 XGBoost 모델의 고장 예측 성능 결과를 비교하였다

  • PDF

Chord-based stepwise Korean Trot music generation technique using RNN-GAN (RNN-GAN을 이용한 코드 기반의 단계적 트로트 음악 생성 기법)

  • Hwang, Seo-Rim;Park, Young-Cheol
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.6
    • /
    • pp.622-628
    • /
    • 2020
  • This paper proposes a music generation technique that automatically generates trot music using a Generative Adversarial Network (GAN) model composed of a Recurrent Neural Network (RNN). The proposed method uses a method of creating a chord as a skeleton of the music, creating a melody and bass in stages based on the chord progression made, and attaching it to the corresponding chord to complete the structured piece. Also, a new chorus chord progression is created from the verse chord progression by applying the characteristics of a trot song that repeats the structure divided into an individual section, such as intro, verse, and chorus. And it extends the length of the created trot. The quality of the generated music was specified using subjective evaluation and objective evaluation methods. It was confirmed that the generated music has similar characteristics to the existing trot.

Improvement of Classification Accuracy of Different Finger Movements Using Surface Electromyography Based on Long Short-Term Memory (LSTM을 이용한 표면 근전도 분석을 통한 서로 다른 손가락 움직임 분류 정확도 향상)

  • Shin, Jaeyoung;Kim, Seong-Uk;Lee, Yun-Sung;Lee, Hyung-Tak;Hwang, Han-Jeong
    • Journal of Biomedical Engineering Research
    • /
    • v.40 no.6
    • /
    • pp.242-249
    • /
    • 2019
  • Forearm electromyography (EMG) generated by wrist movements has been widely used to develop an electrical prosthetic hand, but EMG generated by finger movements has been rarely used even though 20% of amputees lose fingers. The goal of this study is to improve the classification performance of different finger movements using a deep learning algorithm, and thereby contributing to the development of a high-performance finger-based prosthetic hand. Ten participants took part in this study, and they performed seven different finger movements forty times each (thumb, index, middle, ring, little, fist and rest) during which EMG was measured from the back of the right hand using four bipolar electrodes. We extracted mean absolute value (MAV), root mean square (RMS), and mean (MEAN) from the measured EMGs for each trial as features, and a 5x5-fold cross-validation was performed to estimate the classification performance of seven different finger movements. A long short-term memory (LSTM) model was used as a classifier, and linear discriminant analysis (LDA) that is a widely used classifier in previous studies was also used for comparison. The best performance of the LSTM model (sensitivity: 91.46 ± 6.72%; specificity: 91.27 ± 4.18%; accuracy: 91.26 ± 4.09%) significantly outperformed that of LDA (sensitivity: 84.55 ± 9.61%; specificity: 84.02 ± 6.00%; accuracy: 84.00 ± 5.87%). Our result demonstrates the feasibility of a deep learning algorithm (LSTM) to improve the performance of classifying different finger movements using EMG.

A Comparative Study of Machine Learning Algorithms Using LID-DS DataSet (LID-DS 데이터 세트를 사용한 기계학습 알고리즘 비교 연구)

  • Park, DaeKyeong;Ryu, KyungJoon;Shin, DongIl;Shin, DongKyoo;Park, JeongChan;Kim, JinGoog
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.10 no.3
    • /
    • pp.91-98
    • /
    • 2021
  • Today's information and communication technology is rapidly developing, the security of IT infrastructure is becoming more important, and at the same time, cyber attacks of various forms are becoming more advanced and sophisticated like intelligent persistent attacks (Advanced Persistent Threat). Early defense or prediction of increasingly sophisticated cyber attacks is extremely important, and in many cases, the analysis of network-based intrusion detection systems (NIDS) related data alone cannot prevent rapidly changing cyber attacks. Therefore, we are currently using data generated by intrusion detection systems to protect against cyber attacks described above through Host-based Intrusion Detection System (HIDS) data analysis. In this paper, we conducted a comparative study on machine learning algorithms using LID-DS (Leipzig Intrusion Detection-Data Set) host-based intrusion detection data including thread information, metadata, and buffer data missing from previously used data sets. The algorithms used were Decision Tree, Naive Bayes, MLP (Multi-Layer Perceptron), Logistic Regression, LSTM (Long Short-Term Memory model), and RNN (Recurrent Neural Network). Accuracy, accuracy, recall, F1-Score indicators and error rates were measured for evaluation. As a result, the LSTM algorithm had the highest accuracy.

Adaptive Learning Control fo rUnknown Monlinear Systems by Combining Neuro Control and Iterative Learning Control (뉴로제어 및 반복학습제어 기법을 결합한 미지 비선형시스템의 적응학습제어)

  • 최진영;박현주
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.8 no.3
    • /
    • pp.9-15
    • /
    • 1998
  • This paper presents an adaptive learning control method for unknown nonlinear systems by combining neuro control and iterative learning control techniques. In the present control system, an iterative learning controller (ILC) is used for a process of short term memory involved in a temporary adaptive and learning manipulation and a short term storage of a specific temporary action. The learning gain of the iterative learning law is estimated by using a neural network for an unknown system except relative degrees. The control informations obtained by ILC are transferred to a long term memory-based feedforward neuro controller (FNC) and accumulated in it in addition to the previously stored infonnations. This scheme is applied to a two link robot manipulator through simulations.

  • PDF

Rainfall Forecasting Using Satellite Information and Integrated Flood Runoff and Inundation Analysis (I): Theory and Development of Model (위성정보에 의한 강우예측과 홍수유출 및 범람 연계 해석 (I): 이론 및 모형의 개발)

  • Choi, Hyuk Joon;Han, Kun Yeun;Kim, Gwangseob
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.26 no.6B
    • /
    • pp.597-603
    • /
    • 2006
  • The purpose of this study is to improve the short term rainfall forecast skill using neural network model that can deal with the non-linear behavior between satellite data and ground observation, and minimize the flood damage. To overcome the geographical limitation of Korean peninsula and get the long forecast lead time of 3 to 6 hour, the developed rainfall forecast model took satellite imageries and wide range AWS data. The architecture of neural network model is a multi-layer neural network which consists of one input layer, one hidden layer, and one output layer. Neural network is trained using a momentum back propagation algorithm. Flood was estimated using rainfall forecasts. We developed a dynamic flood inundation model which is associated with 1-dimensional flood routing model. Therefore the model can forecast flood aspect in a protected lowland by levee failure of river. In the case of multiple levee breaks at main stream and tributaries, the developed flood inundation model can estimate flood level in a river and inundation level and area in a protected lowland simultaneously.

Estimation of Greenhouse Tomato Transpiration through Mathematical and Deep Neural Network Models Learned from Lysimeter Data (라이시미터 데이터로 학습한 수학적 및 심층 신경망 모델을 통한 온실 토마토 증산량 추정)

  • Meanne P. Andes;Mi-young Roh;Mi Young Lim;Gyeong-Lee Choi;Jung Su Jung;Dongpil Kim
    • Journal of Bio-Environment Control
    • /
    • v.32 no.4
    • /
    • pp.384-395
    • /
    • 2023
  • Since transpiration plays a key role in optimal irrigation management, knowledge of the irrigation demand of crops like tomatoes, which are highly susceptible to water stress, is necessary. One way to determine irrigation demand is to measure transpiration, which is affected by environmental factor or growth stage. This study aimed to estimate the transpiration amount of tomatoes and find a suitable model using mathematical and deep learning models using minute-by-minute data. Pearson correlation revealed that observed environmental variables significantly correlate with crop transpiration. Inside air temperature and outside radiation positively correlated with transpiration, while humidity showed a negative correlation. Multiple Linear Regression (MLR), Polynomial Regression model, Artificial Neural Network (ANN), Long short-term Memory (LSTM), and Gated Recurrent Unit (GRU) models were built and compared their accuracies. All models showed potential in estimating transpiration with R2 values ranging from 0.770 to 0.948 and RMSE of 0.495 mm/min to 1.038 mm/min in the test dataset. Deep learning models outperformed the mathematical models; the GRU demonstrated the best performance in the test data with 0.948 R2 and 0.495 mm/min RMSE. The LSTM and ANN closely followed with R2 values of 0.946 and 0.944, respectively, and RMSE of 0.504 m/min and 0.511, respectively. The GRU model exhibited superior performance in short-term forecasts while LSTM for long-term but requires verification using a large dataset. Compared to the FAO56 Penman-Monteith (PM) equation, PM has a lower RMSE of 0.598 mm/min than MLR and Polynomial models degrees 2 and 3 but performed least among all models in capturing variability in transpiration. Therefore, this study recommended GRU and LSTM models for short-term estimation of tomato transpiration in greenhouses.

A Neural Network for Long-Term Forecast of Regional Precipitation (지역별 중장기 강수량 예측을 위한 신경망 기법)

  • Kim, Ho-Joon;Paek, Hee-Jeong;Kwon, Won-Tae
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.2 no.2
    • /
    • pp.69-78
    • /
    • 1999
  • In this paper, a neural network approach to forecast Korean regional precipitation is presented. We first analyze the characteristics of the conventional models for time series prediction, and then propose a new model and its learning method for the precipitation forecast. The proposed model is a layered network in which the outputs of a layer are buffered within a given period time and then fed fully connected to the upper layer. This study adopted the dual connections between two layers for the model. The network behavior and learning algorithm for the model are also described. The dual connection structure plays the role of the bias of the ordinary Multi-Layer Perceptron(MLP), and reflects the relationships among the features effectively. From these advantageous features, the model provides the learning efficiency in comparison with the FIR network, which is the most popular model for time series prediction. We have applied the model to the monthly and seasonal forecast of precipitation. The precipitation data and SST(Sea Surface Temperature) data for several decades are used as the learning pattern for the neural network predictor. The experimental results have shown the validity of the proposed model.

  • PDF

Symbolizing Numbers to Improve Neural Machine Translation (숫자 기호화를 통한 신경기계번역 성능 향상)

  • Kang, Cheongwoong;Ro, Youngheon;Kim, Jisu;Choi, Heeyoul
    • Journal of Digital Contents Society
    • /
    • v.19 no.6
    • /
    • pp.1161-1167
    • /
    • 2018
  • The development of machine learning has enabled machines to perform delicate tasks that only humans could do, and thus many companies have introduced machine learning based translators. Existing translators have good performances but they have problems in number translation. The translators often mistranslate numbers when the input sentence includes a large number. Furthermore, the output sentence structure completely changes even if only one number in the input sentence changes. In this paper, first, we optimized a neural machine translation model architecture that uses bidirectional RNN, LSTM, and the attention mechanism through data cleansing and changing the dictionary size. Then, we implemented a number-processing algorithm specialized in number translation and applied it to the neural machine translation model to solve the problems above. The paper includes the data cleansing method, an optimal dictionary size and the number-processing algorithm, as well as experiment results for translation performance based on the BLEU score.

A Comparative Study of Machine Learning Algorithms Based on Tensorflow for Data Prediction (데이터 예측을 위한 텐서플로우 기반 기계학습 알고리즘 비교 연구)

  • Abbas, Qalab E.;Jang, Sung-Bong
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.10 no.3
    • /
    • pp.71-80
    • /
    • 2021
  • The selection of an appropriate neural network algorithm is an important step for accurate data prediction in machine learning. Many algorithms based on basic artificial neural networks have been devised to efficiently predict future data. These networks include deep neural networks (DNNs), recurrent neural networks (RNNs), long short-term memory (LSTM) networks, and gated recurrent unit (GRU) neural networks. Developers face difficulties when choosing among these networks because sufficient information on their performance is unavailable. To alleviate this difficulty, we evaluated the performance of each algorithm by comparing their errors and processing times. Each neural network model was trained using a tax dataset, and the trained model was used for data prediction to compare accuracies among the various algorithms. Furthermore, the effects of activation functions and various optimizers on the performance of the models were analyzed The experimental results show that the GRU and LSTM algorithms yields the lowest prediction error with an average RMSE of 0.12 and an average R2 score of 0.78 and 0.75 respectively, and the basic DNN model achieves the lowest processing time but highest average RMSE of 0.163. Furthermore, the Adam optimizer yields the best performance (with DNN, GRU, and LSTM) in terms of error and the worst performance in terms of processing time. The findings of this study are thus expected to be useful for scientists and developers.