• 제목/요약/키워드: Bidirectional Long Short Term Memory (Bi-LSTM)

검색결과 38건 처리시간 0.024초

Assessment of maximum liquefaction distance using soft computing approaches

  • Kishan Kumar;Pijush Samui;Shiva S. Choudhary
    • Geomechanics and Engineering
    • /
    • 제37권4호
    • /
    • pp.395-418
    • /
    • 2024
  • The epicentral region of earthquakes is typically where liquefaction-related damage takes place. To determine the maximum distance, such as maximum epicentral distance (Re), maximum fault distance (Rf), or maximum hypocentral distance (Rh), at which an earthquake can inflict damage, given its magnitude, this study, using a recently updated global liquefaction database, multiple ML models are built to predict the limiting distances (Re, Rf, or Rh) required for an earthquake of a given magnitude to cause damage. Four machine learning models LSTM (Long Short-Term Memory), BiLSTM (Bidirectional Long Short-Term Memory), CNN (Convolutional Neural Network), and XGB (Extreme Gradient Boosting) are developed using the Python programming language. All four proposed ML models performed better than empirical models for limiting distance assessment. Among these models, the XGB model outperformed all the models. In order to determine how well the suggested models can predict limiting distances, a number of statistical parameters have been studied. To compare the accuracy of the proposed models, rank analysis, error matrix, and Taylor diagram have been developed. The ML models proposed in this paper are more robust than other current models and may be used to assess the minimal energy of a liquefaction disaster caused by an earthquake or to estimate the maximum distance of a liquefied site provided an earthquake in rapid disaster mapping.

Neural Model for Named Entity Recognition Considering Aligned Representation

  • Sun, Hongyang;Kim, Taewhan
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2018년도 추계학술발표대회
    • /
    • pp.613-616
    • /
    • 2018
  • Sequence tagging is an important task in Natural Language Processing (NLP), in which the Named Entity Recognition (NER) is the key issue. So far the most widely adopted model for NER in NLP is that of combining the neural network of bidirectional long short-term memory (BiLSTM) and the statistical sequence prediction method of Conditional Random Field (CRF). In this work, we improve the prediction accuracy of the BiLSTM by supporting an aligned word representation mechanism. We have performed experiments on multilingual (English, Spanish and Dutch) datasets and confirmed that our proposed model outperformed the existing state-of-the-art models.

Video Saliency Detection Using Bi-directional LSTM

  • Chi, Yang;Li, Jinjiang
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제14권6호
    • /
    • pp.2444-2463
    • /
    • 2020
  • Significant detection of video can more rationally allocate computing resources and reduce the amount of computation to improve accuracy. Deep learning can extract the edge features of the image, providing technical support for video saliency. This paper proposes a new detection method. We combine the Convolutional Neural Network (CNN) and the Deep Bidirectional LSTM Network (DB-LSTM) to learn the spatio-temporal features by exploring the object motion information and object motion information to generate video. A continuous frame of significant images. We also analyzed the sample database and found that human attention and significant conversion are time-dependent, so we also considered the significance detection of video cross-frame. Finally, experiments show that our method is superior to other advanced methods.

재귀 신경망에 기반을 둔 트래픽 부하 예측을 이용한 적응적 안테나 뮤팅 (Adaptive Antenna Muting using RNN-based Traffic Load Prediction)

  • Ahmadzai, Fazel Haq;Lee, Woongsup
    • 한국정보통신학회논문지
    • /
    • 제26권4호
    • /
    • pp.633-636
    • /
    • 2022
  • The reduction of energy consumption at the base station (BS) has become more important recently. In this paper, we consider the adaptive muting of the antennas based on the predicted future traffic load to reduce the energy consumption where the number of active antennas is adaptively adjusted according to the predicted future traffic load. Given that traffic load is sequential data, three different RNN structures, namely long-short term memory (LSTM), gated recurrent unit (GRU), and bidirectional LSTM (Bi-LSTM) are considered for the future traffic load prediction. Through the performance evaluation based on the actual traffic load collected from the Afghanistan telecom company, we confirm that the traffic load can be estimated accurately and the overall power consumption can also be reduced significantly using the antenna musing.

BiLSTM 기반의 설명 가능한 태양광 발전량 예측 기법 (Explainable Photovoltaic Power Forecasting Scheme Using BiLSTM)

  • 박성우;정승민;문재욱;황인준
    • 정보처리학회논문지:소프트웨어 및 데이터공학
    • /
    • 제11권8호
    • /
    • pp.339-346
    • /
    • 2022
  • 최근 화석연료의 무분별한 사용으로 인한 자원고갈 문제 및 기후변화 문제 등이 심각해짐에 따라 화석연료를 대체할 수 있는 신재생에너지에 대한 관심이 증가하고 있다. 특히 신재생에너지 중 태양광 에너지는 다른 신재생에너지원에 비해 고갈될 염려가 적고, 공간적인 제약이 크지 않아 전국적으로 수요가 증가하고 있다. 태양광 발전 시스템에서 생산된 전력을 효율적으로 사용하기 위해서는 보다 정확한 태양광 발전량 예측 모델이 필요하다. 이를 위하여 다양한 기계학습 및 심층학습 기반의 태양광 발전량 예측 모델이 제안되었지만, 심층학습 기반의 예측 모델은 모델 내부에서 일어나는 의사결정 과정을 해석하기가 어렵다는 단점을 보유하고 있다. 이러한 문제를 해결하기 위하여 설명 가능한 인공지능 기술이 많은 주목을 받고 있다. 설명 가능한 인공지능 기술을 통하여 예측 모델의 결과 도출 과정을 해석할 수 있다면 모델의 신뢰성을 확보할 수 있을 뿐만 아니라 해석된 도출 결과를 바탕으로 모델을 개선하여 성능 향상을 기대할 수도 있다. 이에 본 논문에서는 BiLSTM(Bidirectional Long Short-Term Memory)을 사용하여 모델을 구성하고, 모델에서 어떻게 예측값이 도출되었는지를 SHAP(SHapley Additive exPlanations)을 통하여 설명하는 설명 가능한 태양광 발전량 예측 기법을 제안한다.

Bi-LSTM model with time distribution for bandwidth prediction in mobile networks

  • Hyeonji Lee;Yoohwa Kang;Minju Gwak;Donghyeok An
    • ETRI Journal
    • /
    • 제46권2호
    • /
    • pp.205-217
    • /
    • 2024
  • We propose a bandwidth prediction approach based on deep learning. The approach is intended to accurately predict the bandwidth of various types of mobile networks. We first use a machine learning technique, namely, the gradient boosting algorithm, to recognize the connected mobile network. Second, we apply a handover detection algorithm based on network recognition to account for vertical handover that causes the bandwidth variance. Third, as the communication performance offered by 3G, 4G, and 5G networks varies, we suggest a bidirectional long short-term memory model with time distribution for bandwidth prediction per network. To increase the prediction accuracy, pretraining and fine-tuning are applied for each type of network. We use a dataset collected at University College Cork for network recognition, handover detection, and bandwidth prediction. The performance evaluation indicates that the handover detection algorithm achieves 88.5% accuracy, and the bandwidth prediction model achieves a high accuracy, with a root-mean-square error of only 2.12%.

A Deep Learning Model for Extracting Consumer Sentiments using Recurrent Neural Network Techniques

  • Ranjan, Roop;Daniel, AK
    • International Journal of Computer Science & Network Security
    • /
    • 제21권8호
    • /
    • pp.238-246
    • /
    • 2021
  • The rapid rise of the Internet and social media has resulted in a large number of text-based reviews being placed on sites such as social media. In the age of social media, utilizing machine learning technologies to analyze the emotional context of comments aids in the understanding of QoS for any product or service. The classification and analysis of user reviews aids in the improvement of QoS. (Quality of Services). Machine Learning algorithms have evolved into a powerful tool for analyzing user sentiment. Unlike traditional categorization models, which are based on a set of rules. In sentiment categorization, Bidirectional Long Short-Term Memory (BiLSTM) has shown significant results, and Convolution Neural Network (CNN) has shown promising results. Using convolutions and pooling layers, CNN can successfully extract local information. BiLSTM uses dual LSTM orientations to increase the amount of background knowledge available to deep learning models. The suggested hybrid model combines the benefits of these two deep learning-based algorithms. The data source for analysis and classification was user reviews of Indian Railway Services on Twitter. The suggested hybrid model uses the Keras Embedding technique as an input source. The suggested model takes in data and generates lower-dimensional characteristics that result in a categorization result. The suggested hybrid model's performance was compared using Keras and Word2Vec, and the proposed model showed a significant improvement in response with an accuracy of 95.19 percent.

Point of Interest Recommendation System Using Sentiment Analysis

  • Gaurav Meena;Ajay Indian;Krishna Kumar Mohbey;Kunal Jangid
    • Journal of Information Science Theory and Practice
    • /
    • 제12권2호
    • /
    • pp.64-78
    • /
    • 2024
  • Sentiment analysis is one of the promising approaches for developing a point of interest (POI) recommendation system. It uses natural language processing techniques that deploy expert insights from user-generated content such as reviews and feedback. By applying sentiment polarities (positive, negative, or neutral) associated with each POI, the recommendation system can suggest the most suitable POIs for specific users. The proposed study combines two models for POI recommendation. The first model uses bidirectional long short-term memory (BiLSTM) to predict sentiments and is trained on an election dataset. It is observed that the proposed model outperforms existing models in terms of accuracy (99.52%), precision (99.53%), recall (99.51%), and F1-score (99.52%). Then, this model is used on the Foursquare dataset to predict the class labels. Following this, user and POI embeddings are generated. The next model recommends the top POIs and corresponding coordinates to the user using the LSTM model. Filtered user interest and locations are used to recommend POIs from the Foursquare dataset. The results of our proposed model for the POI recommendation system using sentiment analysis are compared to several state-of-the-art approaches and are found quite affirmative regarding recall (48.5%) and precision (85%). The proposed system can be used for trip advice, group recommendations, and interesting place recommendations to specific users.

RDNN: Rumor Detection Neural Network for Veracity Analysis in Social Media Text

  • SuthanthiraDevi, P;Karthika, S
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제16권12호
    • /
    • pp.3868-3888
    • /
    • 2022
  • A widely used social networking service like Twitter has the ability to disseminate information to large groups of people even during a pandemic. At the same time, it is a convenient medium to share irrelevant and unverified information online and poses a potential threat to society. In this research, conventional machine learning algorithms are analyzed to classify the data as either non-rumor data or rumor data. Machine learning techniques have limited tuning capability and make decisions based on their learning. To tackle this problem the authors propose a deep learning-based Rumor Detection Neural Network model to predict the rumor tweet in real-world events. This model comprises three layers, AttCNN layer is used to extract local and position invariant features from the data, AttBi-LSTM layer to extract important semantic or contextual information and HPOOL to combine the down sampling patches of the input feature maps from the average and maximum pooling layers. A dataset from Kaggle and ground dataset #gaja are used to train the proposed Rumor Detection Neural Network to determine the veracity of the rumor. The experimental results of the RDNN Classifier demonstrate an accuracy of 93.24% and 95.41% in identifying rumor tweets in real-time events.

A Network Intrusion Security Detection Method Using BiLSTM-CNN in Big Data Environment

  • Hong Wang
    • Journal of Information Processing Systems
    • /
    • 제19권5호
    • /
    • pp.688-701
    • /
    • 2023
  • The conventional methods of network intrusion detection system (NIDS) cannot measure the trend of intrusiondetection targets effectively, which lead to low detection accuracy. In this study, a NIDS method which based on a deep neural network in a big-data environment is proposed. Firstly, the entire framework of the NIDS model is constructed in two stages. Feature reduction and anomaly probability output are used at the core of the two stages. Subsequently, a convolutional neural network, which encompasses a down sampling layer and a characteristic extractor consist of a convolution layer, the correlation of inputs is realized by introducing bidirectional long short-term memory. Finally, after the convolution layer, a pooling layer is added to sample the required features according to different sampling rules, which promotes the overall performance of the NIDS model. The proposed NIDS method and three other methods are compared, and it is broken down under the conditions of the two databases through simulation experiments. The results demonstrate that the proposed model is superior to the other three methods of NIDS in two databases, in terms of precision, accuracy, F1- score, and recall, which are 91.64%, 93.35%, 92.25%, and 91.87%, respectively. The proposed algorithm is significant for improving the accuracy of NIDS.