• Title/Summary/Keyword: long term neural network

Search Result 395, Processing Time 0.03 seconds

Analyzing the Impact of Multivariate Inputs on Deep Learning-Based Reservoir Level Prediction and Approaches for Mid to Long-Term Forecasting (다변량 입력이 딥러닝 기반 저수율 예측에 미치는 영향 분석과 중장기 예측 방안)

  • Hyeseung Park;Jongwook Yoon;Hojun Lee;Hyunho Yang
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.4
    • /
    • pp.199-207
    • /
    • 2024
  • Local reservoirs are crucial sources for agricultural water supply, necessitating stable water level management to prepare for extreme climate conditions such as droughts. Water level prediction is significantly influenced by local climate characteristics, such as localized rainfall, as well as seasonal factors including cropping times, making it essential to understand the correlation between input and output data as much as selecting an appropriate prediction model. In this study, extensive multivariate data from over 400 reservoirs in Jeollabuk-do from 1991 to 2022 was utilized to train and validate a water level prediction model that comprehensively reflects the complex hydrological and climatological environmental factors of each reservoir, and to analyze the impact of each input feature on the prediction performance of water levels. Instead of focusing on improvements in water level performance through neural network structures, the study adopts a basic Feedforward Neural Network composed of fully connected layers, batch normalization, dropout, and activation functions, focusing on the correlation between multivariate input data and prediction performance. Additionally, most existing studies only present short-term prediction performance on a daily basis, which is not suitable for practical environments that require medium to long-term predictions, such as 10 days or a month. Therefore, this study measured the water level prediction performance up to one month ahead through a recursive method that uses daily prediction values as the next input. The experiment identified performance changes according to the prediction period and analyzed the impact of each input feature on the overall performance based on an Ablation study.

Use of an Artificial Neural Network to Predict Risk Factors of Nosocomial Infection in Lung Cancer Patients

  • Chen, Jie;Pan, Qin-Shi;Hong, Wan-Dong;Pan, Jingye;Zhang, Wen-Hui;Xu, Gang;Wang, Yu-Min
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.15 no.13
    • /
    • pp.5349-5353
    • /
    • 2014
  • Statistical methods to analyze and predict the related risk factors of nosocomial infection in lung cancer patients are various, but the results are inconsistent. A total of 609 patients with lung cancer were enrolled to allow factor comparison using Student's t-test or the Mann-Whitney test or the Chi-square test. Variables that were significantly related to the presence of nosocomial infection were selected as candidates for input into the final ANN model. The area under the receiver operating characteristic (ROC) curve (AUC) was used to evaluate the performance of the artificial neural network (ANN) model and logistic regression (LR) model. The prevalence of nosocomial infection from lung cancer in this entire study population was 20.1% (165/609), nosocomial infections occurring in sputum specimens (85.5%), followed by blood (6.73%), urine (6.0%) and pleural effusions (1.82%). It was shown that long term hospitalization (${\geq}22days$, P= 0.000), poor clinical stage (IIIb and IV stage, P=0.002), older age (${\geq}61days$ old, P=0.023), and use the hormones were linked to nosocomial infection and the ANN model consisted of these four factors. The artificial neural network model with variables consisting of age, clinical stage, time of hospitalization, and use of hormones should be useful for predicting nosocomial infection in lung cancer cases.

Modeling of wind and temperature effects on modal frequencies and analysis of relative strength of effect

  • Zhou, H.F.;Ni, Y.Q.;Ko, J.M.;Wong, K.Y.
    • Wind and Structures
    • /
    • v.11 no.1
    • /
    • pp.35-50
    • /
    • 2008
  • Wind and temperature have been shown to be the critical sources causing changes in the modal properties of large-scale bridges. While the individual effects of wind and temperature on modal variability have been widely studied, the investigation about the effects of multiple environmental factors on structural modal properties was scarcely reported. This paper addresses the modeling of the simultaneous effects of wind and temperature on the modal frequencies of an instrumented cable-stayed bridge. Making use of the long-term monitoring data from anemometers, temperature sensors and accelerometers, a neural network model is formulated to correlate the modal frequency of each vibration mode with wind speed and temperature simultaneously. Research efforts have been made on enhancing the prediction capability of the neural network model through optimal selection of the number of hidden nodes and an analysis of relative strength of effect (RSE) for input reconstruction. The generalization performance of the formulated model is verified with a set of new testing data that have not been used in formulating the model. It is shown that using the significant components of wind speeds and temperatures rather than the whole measurement components as input to neural network can enhance the prediction capability. For the fundamental mode of the bridge investigated, wind and temperature together apply an overall negative action on the modal frequency, and the change in wind condition contributes less to the modal variability than the change in temperature.

A Method for Generating Malware Countermeasure Samples Based on Pixel Attention Mechanism

  • Xiangyu Ma;Yuntao Zhao;Yongxin Feng;Yutao Hu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.2
    • /
    • pp.456-477
    • /
    • 2024
  • With information technology's rapid development, the Internet faces serious security problems. Studies have shown that malware has become a primary means of attacking the Internet. Therefore, adversarial samples have become a vital breakthrough point for studying malware. By studying adversarial samples, we can gain insights into the behavior and characteristics of malware, evaluate the performance of existing detectors in the face of deceptive samples, and help to discover vulnerabilities and improve detection methods for better performance. However, existing adversarial sample generation methods still need help regarding escape effectiveness and mobility. For instance, researchers have attempted to incorporate perturbation methods like Fast Gradient Sign Method (FGSM), Projected Gradient Descent (PGD), and others into adversarial samples to obfuscate detectors. However, these methods are only effective in specific environments and yield limited evasion effectiveness. To solve the above problems, this paper proposes a malware adversarial sample generation method (PixGAN) based on the pixel attention mechanism, which aims to improve adversarial samples' escape effect and mobility. The method transforms malware into grey-scale images and introduces the pixel attention mechanism in the Deep Convolution Generative Adversarial Networks (DCGAN) model to weigh the critical pixels in the grey-scale map, which improves the modeling ability of the generator and discriminator, thus enhancing the escape effect and mobility of the adversarial samples. The escape rate (ASR) is used as an evaluation index of the quality of the adversarial samples. The experimental results show that the adversarial samples generated by PixGAN achieve escape rates of 97%, 94%, 35%, 39%, and 43% on the Random Forest (RF), Support Vector Machine (SVM), Convolutional Neural Network (CNN), Convolutional Neural Network and Recurrent Neural Network (CNN_RNN), and Convolutional Neural Network and Long Short Term Memory (CNN_LSTM) algorithmic detectors, respectively.

Performance Analysis of Deep Learning-based Normalization According to Input-output Structure and Neural Network Model (입출력구조와 신경망 모델에 따른 딥러닝 기반 정규화 기법의 성능 분석)

  • Changsoo Ryu;Geunhwan Kim
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.29 no.4
    • /
    • pp.13-24
    • /
    • 2024
  • In this paper, we analyzed the performance of normalization according to various neural network models and input-output structures. For the analysis, a simulation-based dataset for noise environments with homogeneous and up to three interfering signals was used. As a result, the end-to-end structure that directly outputs noise variance showed superior performance when using a 1-D convolutional neural network and BiLSTM model, and was analyzed to be particularly robust against interference signals. This is because the 1-D convolutional neural network and bidirectional long short-term memory models have stronger inductive bias than the multilayer perceptron and transformer models. The analysis of this paper are expected to be used as a useful reference for future research on deep learning-based normalization.

Context-Aware Mobile User Authentication Approach using LSTM networks (LSTM 신경망을 활용한 맥락 기반 모바일 사용자 인증 기법)

  • Nam, Sangjin;Kim, Suntae;Shin, Jung-Hoon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.1
    • /
    • pp.11-18
    • /
    • 2020
  • This study aims to complement the poor performance of existing context-aware authentication techniques in the mobile environment. The data used are GPS, Call Detail Record(CDR) and app usage. locational classification according to GPS density was implemented in order to distinguish other people in populated areas in the processing of GPS. It also handles missing values that may occur in data collection. The authentication model consists of two long-short term memory(LSTM) and one Artificial Neural Network(ANN) that aggregates the results, which produces authentication scores. In this paper, we compare the accuracy of this technique with that of other studies. Then compare the number of authentication attempts required to detect someone else's authentication. As a result, we achieved an average 11.6% improvement in accuracy and faster detection of approximately 60% of the experimental data.

An LSTM Neural Network Model for Forecasting Daily Peak Electric Load of EV Charging Stations (EV 충전소의 일별 최대전력부하 예측을 위한 LSTM 신경망 모델)

  • Lee, Haesung;Lee, Byungsung;Ahn, Hyun
    • Journal of Internet Computing and Services
    • /
    • v.21 no.5
    • /
    • pp.119-127
    • /
    • 2020
  • As the electric vehicle (EV) market in South Korea grows, it is required to expand charging facilities to respond to rapidly increasing EV charging demand. In order to conduct a comprehensive facility planning, it is necessary to forecast future demand for electricity and systematically analyze the impact on the load capacity of facilities based on this. In this paper, we design and develop a Long Short-Term Memory (LSTM) neural network model that predicts the daily peak electric load at each charging station using the EV charging data of KEPCO. First, we obtain refined data through data preprocessing and outlier removal. Next, our model is trained by extracting daily features per charging station and constructing a training set. Finally, our model is verified through performance analysis using a test set for each charging station type, and the limitations of our model are discussed.

Forecasting of Iron Ore Prices using Machine Learning (머신러닝을 이용한 철광석 가격 예측에 대한 연구)

  • Lee, Woo Chang;Kim, Yang Sok;Kim, Jung Min;Lee, Choong Kwon
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.25 no.2
    • /
    • pp.57-72
    • /
    • 2020
  • The price of iron ore has continued to fluctuate with high demand and supply from many countries and companies. In this business environment, forecasting the price of iron ore has become important. This study developed the machine learning model forecasting the price of iron ore a one month after the trading events. The forecasting model used distributed lag model and deep learning models such as MLP (Multi-layer perceptron), RNN (Recurrent neural network) and LSTM (Long short-term memory). According to the results of comparing individual models through metrics, LSTM showed the lowest predictive error. Also, as a result of comparing the models using the ensemble technique, the distributed lag and LSTM ensemble model showed the lowest prediction.

A Korean speech recognition based on conformer (콘포머 기반 한국어 음성인식)

  • Koo, Myoung-Wan
    • The Journal of the Acoustical Society of Korea
    • /
    • v.40 no.5
    • /
    • pp.488-495
    • /
    • 2021
  • We propose a speech recognition system based on conformer. Conformer is known to be convolution-augmented transformer, which combines transfer model for capturing global information with Convolution Neural Network (CNN) for exploiting local feature effectively. The baseline system is developed to be a transfer-based speech recognition using Long Short-Term Memory (LSTM)-based language model. The proposed system is a system which uses conformer instead of transformer with transformer-based language model. When Electronics and Telecommunications Research Institute (ETRI) speech corpus in AI-Hub is used for our evaluation, the proposed system yields 5.7 % of Character Error Rate (CER) while the baseline system results in 11.8 % of CER. Even though speech corpus is extended into other domain of AI-hub such as NHNdiguest speech corpus, the proposed system makes a robust performance for two domains. Throughout those experiments, we can prove a validation of the proposed system.

Comparison of Power Consumption Prediction Scheme Based on Artificial Intelligence (인공지능 기반 전력량예측 기법의 비교)

  • Lee, Dong-Gu;Sun, Young-Ghyu;Kim, Soo-Hyun;Sim, Issac;Hwang, Yu-Min;Kim, Jin-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.4
    • /
    • pp.161-167
    • /
    • 2019
  • Recently, demand forecasting techniques have been actively studied due to interest in stable power supply with surging power demand, and increase in spread of smart meters that enable real-time power measurement. In this study, we proceeded the deep learning prediction model experiments which learns actual measured power usage data of home and outputs the forecasting result. And we proceeded pre-processing with moving average method. The predicted value made by the model is evaluated with the actual measured data. Through this forecasting, it is possible to lower the power supply reserve ratio and reduce the waste of the unused power. In this paper, we conducted experiments on three types of networks: Multi Layer Perceptron (MLP), Recurrent Neural Network (RNN), and Long Short Term Memory (LSTM) and we evaluate the results of each scheme. Evaluation is conducted with following method: MSE(Mean Squared Error) method and MAE(Mean Absolute Error).