• Title/Summary/Keyword: deep machine learning

Search Result 1,085, Processing Time 0.031 seconds

Hourly Water Level Simulation in Tancheon River Using an LSTM (LSTM을 이용한 탄천에서의 시간별 하천수위 모의)

  • Park, Chang Eon
    • Journal of The Korean Society of Agricultural Engineers
    • /
    • v.66 no.4
    • /
    • pp.51-57
    • /
    • 2024
  • This study was conducted on how to simulate runoff, which was done using existing physical models, using an LSTM (Long Short-Term Memory) model based on deep learning. Tancheon, the first tributary of the Han River, was selected as the target area for the model application. To apply the model, one water level observatory and four rainfall observatories were selected, and hourly data from 2020 to 2023 were collected to apply the model. River water level of the outlet of the Tancheon basin was simulated by inputting precipitation data from four rainfall observation stations in the basin and average preceding 72-hour precipitation data for each hour. As a result of water level simulation using 2021 to 2023 data for learning and testing with 2020 data, it was confirmed that reliable simulation results were produced through appropriate learning steps, reaching a certain mean absolute error in a short period time. Despite the short data period, it was found that the mean absolute percentage error was 0.5544~0.6226%, showing an accuracy of over 99.4%. As a result of comparing the simulated and observed values of the rapidly changing river water level during a specific heavy rain period, the coefficient of determination was found to be 0.9754 and 0.9884. It was determined that the performance of LSTM, which aims to simulate river water levels, could be improved by including preceding precipitation in the input data and using precipitation data from various rainfall observation stations within the basin.

A Time Series Graph based Convolutional Neural Network Model for Effective Input Variable Pattern Learning : Application to the Prediction of Stock Market (효과적인 입력변수 패턴 학습을 위한 시계열 그래프 기반 합성곱 신경망 모형: 주식시장 예측에의 응용)

  • Lee, Mo-Se;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.167-181
    • /
    • 2018
  • Over the past decade, deep learning has been in spotlight among various machine learning algorithms. In particular, CNN(Convolutional Neural Network), which is known as the effective solution for recognizing and classifying images or voices, has been popularly applied to classification and prediction problems. In this study, we investigate the way to apply CNN in business problem solving. Specifically, this study propose to apply CNN to stock market prediction, one of the most challenging tasks in the machine learning research. As mentioned, CNN has strength in interpreting images. Thus, the model proposed in this study adopts CNN as the binary classifier that predicts stock market direction (upward or downward) by using time series graphs as its inputs. That is, our proposal is to build a machine learning algorithm that mimics an experts called 'technical analysts' who examine the graph of past price movement, and predict future financial price movements. Our proposed model named 'CNN-FG(Convolutional Neural Network using Fluctuation Graph)' consists of five steps. In the first step, it divides the dataset into the intervals of 5 days. And then, it creates time series graphs for the divided dataset in step 2. The size of the image in which the graph is drawn is $40(pixels){\times}40(pixels)$, and the graph of each independent variable was drawn using different colors. In step 3, the model converts the images into the matrices. Each image is converted into the combination of three matrices in order to express the value of the color using R(red), G(green), and B(blue) scale. In the next step, it splits the dataset of the graph images into training and validation datasets. We used 80% of the total dataset as the training dataset, and the remaining 20% as the validation dataset. And then, CNN classifiers are trained using the images of training dataset in the final step. Regarding the parameters of CNN-FG, we adopted two convolution filters ($5{\times}5{\times}6$ and $5{\times}5{\times}9$) in the convolution layer. In the pooling layer, $2{\times}2$ max pooling filter was used. The numbers of the nodes in two hidden layers were set to, respectively, 900 and 32, and the number of the nodes in the output layer was set to 2(one is for the prediction of upward trend, and the other one is for downward trend). Activation functions for the convolution layer and the hidden layer were set to ReLU(Rectified Linear Unit), and one for the output layer set to Softmax function. To validate our model - CNN-FG, we applied it to the prediction of KOSPI200 for 2,026 days in eight years (from 2009 to 2016). To match the proportions of the two groups in the independent variable (i.e. tomorrow's stock market movement), we selected 1,950 samples by applying random sampling. Finally, we built the training dataset using 80% of the total dataset (1,560 samples), and the validation dataset using 20% (390 samples). The dependent variables of the experimental dataset included twelve technical indicators popularly been used in the previous studies. They include Stochastic %K, Stochastic %D, Momentum, ROC(rate of change), LW %R(Larry William's %R), A/D oscillator(accumulation/distribution oscillator), OSCP(price oscillator), CCI(commodity channel index), and so on. To confirm the superiority of CNN-FG, we compared its prediction accuracy with the ones of other classification models. Experimental results showed that CNN-FG outperforms LOGIT(logistic regression), ANN(artificial neural network), and SVM(support vector machine) with the statistical significance. These empirical results imply that converting time series business data into graphs and building CNN-based classification models using these graphs can be effective from the perspective of prediction accuracy. Thus, this paper sheds a light on how to apply deep learning techniques to the domain of business problem solving.

Voice Activity Detection based on DBN using the Likelihood Ratio (우도비를 이용한 DBN 기반의 음성 검출기)

  • Kim, S.K.;Lee, S.M.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.8 no.3
    • /
    • pp.145-150
    • /
    • 2014
  • In this paper, we propose a novel scheme to improve the performance of a voice activity detection(VAD) which is based on the deep belief networks(DBN) with the likelihood ratio(LR). The proposed algorithm applies the DBN learning method which is trained in order to minimize the probability of detection error instead of the conventional decision rule using geometric mean. Experimental results show that the proposed algorithm yields better results compared to the conventional VAD algorithm in various noise environments.

  • PDF

River Water Level Prediction Method based on LSTM Neural Network

  • Le, Xuan Hien;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2018.05a
    • /
    • pp.147-147
    • /
    • 2018
  • In this article, we use an open source software library: TensorFlow, developed for the purposes of conducting very complex machine learning and deep neural network applications. However, the system is general enough to be applicable in a wide variety of other domains as well. The proposed model based on a deep neural network model, LSTM (Long Short-Term Memory) to predict the river water level at Okcheon Station of the Guem River without utilization of rainfall - forecast information. For LSTM modeling, the input data is hourly water level data for 15 years from 2002 to 2016 at 4 stations includes 3 upstream stations (Sutong, Hotan, and Songcheon) and the forecasting-target station (Okcheon). The data are subdivided into three purposes: a training data set, a testing data set and a validation data set. The model was formulated to predict Okcheon Station water level for many cases from 3 hours to 12 hours of lead time. Although the model does not require many input data such as climate, geography, land-use for rainfall-runoff simulation, the prediction is very stable and reliable up to 9 hours of lead time with the Nash - Sutcliffe efficiency (NSE) is higher than 0.90 and the root mean square error (RMSE) is lower than 12cm. The result indicated that the method is able to produce the river water level time series and be applicable to the practical flood forecasting instead of hydrologic modeling approaches.

  • PDF

Detonation cell size model based on deep neural network for hydrogen, methane and propane mixtures with air and oxygen

  • Malik, Konrad;Zbikowski, Mateusz;Teodorczyk, Andrzej
    • Nuclear Engineering and Technology
    • /
    • v.51 no.2
    • /
    • pp.424-431
    • /
    • 2019
  • The aim of the present study was to develop model for detonation cell sizes prediction based on a deep artificial neural network of hydrogen, methane and propane mixtures with air and oxygen. The discussion about the currently available algorithms compared existing solutions and resulted in a conclusion that there is a need for a new model, free from uncertainty of the effective activation energy and the reaction length definitions. The model offers a better and more feasible alternative to the existing ones. Resulting predictions were validated against experimental data obtained during the investigation of detonation parameters, as well as with data collected from the literature. Additionally, separate models for individual mixtures were created and compared with the main model. The comparison showed no drawbacks caused by fitting one model to many mixtures. Moreover, it was demonstrated that the model may be easily extended by including more independent variables. As an example, dependency on pressure was examined. The preparation of experimental data for deep neural network training was described in detail to allow reproducing the results obtained and extending the model to different mixtures and initial conditions. The source code of ready to use models is also provided.

A Video Expression Recognition Method Based on Multi-mode Convolution Neural Network and Multiplicative Feature Fusion

  • Ren, Qun
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.556-570
    • /
    • 2021
  • The existing video expression recognition methods mainly focus on the spatial feature extraction of video expression images, but tend to ignore the dynamic features of video sequences. To solve this problem, a multi-mode convolution neural network method is proposed to effectively improve the performance of facial expression recognition in video. Firstly, OpenFace 2.0 is used to detect face images in video, and two deep convolution neural networks are used to extract spatiotemporal expression features. Furthermore, spatial convolution neural network is used to extract the spatial information features of each static expression image, and the dynamic information feature is extracted from the optical flow information of multiple expression images based on temporal convolution neural network. Then, the spatiotemporal features learned by the two deep convolution neural networks are fused by multiplication. Finally, the fused features are input into support vector machine to realize the facial expression classification. Experimental results show that the recognition accuracy of the proposed method can reach 64.57% and 60.89%, respectively on RML and Baum-ls datasets. It is better than that of other contrast methods.

A Novel SOC Estimation Method for Multiple Number of Lithium Batteries Using a Deep Neural Network (딥 뉴럴 네트워크를 이용한 새로운 리튬이온 배터리의 SOC 추정법)

  • Khan, Asad;Ko, Young-Hwi;Choi, Woo-Jin
    • The Transactions of the Korean Institute of Power Electronics
    • /
    • v.26 no.1
    • /
    • pp.1-8
    • /
    • 2021
  • For the safe and reliable operation of lithium-ion batteries in electric vehicles or energy storage systems, having accurate information of the battery, such as the state of charge (SOC), is essential. Many different techniques of battery SOC estimation have been developed, such as the Kalman filter. However, when this filter is applied to multiple batteries, it has difficulty maintaining the accuracy of the estimation over all cells owing to the difference in parameter values of each cell. The difference in the parameter of each cell may increase as the operation time accumulates due to aging. In this paper, a novel deep neural network (DNN)-based SOC estimation method for multi-cell application is proposed. In the proposed method, DNN is implemented to determine the nonlinear relationships of the voltage and current at different SOCs and temperatures. In the training, the voltage and current data obtained at different temperatures during charge/discharge cycles are used. After the comprehensive training with the data obtained from the cycle test with a cell, the resulting algorithm is applied to estimate the SOC of other cells. Experimental results show that the mean absolute error of the estimation is 1.213% at 25℃ with the proposed DNN-based SOC estimation method.

Application of machine learning and deep neural network for wave propagation in lung cancer cell

  • Xing, Lumin;Liu, Wenjian;Li, Xin;Wang, Han;Jiang, Zhiming;Wang, Lingling
    • Advances in nano research
    • /
    • v.13 no.3
    • /
    • pp.297-312
    • /
    • 2022
  • Coughing and breath shortness are common symptoms of nano (small) cell lung cancer. Smoking is main factor in causing such cancers. The cancer cells form on the soft tissues of lung. Deformation behavior and wave vibration of lung affected when cancer cells exist. Therefore, in the current work, phase velocity behavior of the small cell lung cancer as a main part of the body via an exact size-dependent theory is presented. Regarding this problem, displacement fields of small cell lung cancer are obtained using first-order shear deformation theory with five parameters. Besides, the size-dependent small cell lung cancer is modeled via nonlocal stress/strain gradient theory (NSGT). An analytical method is applied for solving the governing equations of the small cell lung cancer structure. The novelty of the current study is the consideration of the five-parameter of displacement for curved panel, and porosity as well as NSGT are employed and solved using the analytical method. For more verification, the outcomes of this reports are compared with the predictions of deep neural network (DNN) with adaptive optimization method. A thorough parametric investigation is conducted on the effect of NSGT parameters, porosity and geometry on the phase velocity behavior of the small cell lung cancer structure.

Counterfactual image generation by disentangling data attributes with deep generative models

  • Jieon Lim;Weonyoung Joo
    • Communications for Statistical Applications and Methods
    • /
    • v.30 no.6
    • /
    • pp.589-603
    • /
    • 2023
  • Deep generative models target to infer the underlying true data distribution, and it leads to a huge success in generating fake-but-realistic data. Regarding such a perspective, the data attributes can be a crucial factor in the data generation process since non-existent counterfactual samples can be generated by altering certain factors. For example, we can generate new portrait images by flipping the gender attribute or altering the hair color attributes. This paper proposes counterfactual disentangled variational autoencoder generative adversarial networks (CDVAE-GAN), specialized for data attribute level counterfactual data generation. The structure of the proposed CDVAE-GAN consists of variational autoencoders and generative adversarial networks. Specifically, we adopt a Gaussian variational autoencoder to extract low-dimensional disentangled data features and auxiliary Bernoulli latent variables to model the data attributes separately. Also, we utilize a generative adversarial network to generate data with high fidelity. By enjoying the benefits of the variational autoencoder with the additional Bernoulli latent variables and the generative adversarial network, the proposed CDVAE-GAN can control the data attributes, and it enables producing counterfactual data. Our experimental result on the CelebA dataset qualitatively shows that the generated samples from CDVAE-GAN are realistic. Also, the quantitative results support that the proposed model can produce data that can deceive other machine learning classifiers with the altered data attributes.

Detecting Adversarial Example Using Ensemble Method on Deep Neural Network (딥뉴럴네트워크에서의 적대적 샘플에 관한 앙상블 방어 연구)

  • Kwon, Hyun;Yoon, Joonhyeok;Kim, Junseob;Park, Sangjun;Kim, Yongchul
    • Convergence Security Journal
    • /
    • v.21 no.2
    • /
    • pp.57-66
    • /
    • 2021
  • Deep neural networks (DNNs) provide excellent performance for image, speech, and pattern recognition. However, DNNs sometimes misrecognize certain adversarial examples. An adversarial example is a sample that adds optimized noise to the original data, which makes the DNN erroneously misclassified, although there is nothing wrong with the human eye. Therefore studies on defense against adversarial example attacks are required. In this paper, we have experimentally analyzed the success rate of detection for adversarial examples by adjusting various parameters. The performance of the ensemble defense method was analyzed using fast gradient sign method, DeepFool method, Carlini & Wanger method, which are adversarial example attack methods. Moreover, we used MNIST as experimental data and Tensorflow as a machine learning library. As an experimental method, we carried out performance analysis based on three adversarial example attack methods, threshold, number of models, and random noise. As a result, when there were 7 models and a threshold of 1, the detection rate for adversarial example is 98.3%, and the accuracy of 99.2% of the original sample is maintained.