• Title/Summary/Keyword: neural network optimization

Search Result 816, Processing Time 0.024 seconds

Implementation of Unsupervised Nonlinear Classifier with Binary Harmony Search Algorithm (Binary Harmony Search 알고리즘을 이용한 Unsupervised Nonlinear Classifier 구현)

  • Lee, Tae-Ju;Park, Seung-Min;Ko, Kwang-Eun;Sung, Won-Ki;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.23 no.4
    • /
    • pp.354-359
    • /
    • 2013
  • In this paper, we suggested the method for implementation of unsupervised nonlinear classification using Binary Harmony Search (BHS) algorithm, which is known as a optimization algorithm. Various algorithms have been suggested for classification of feature vectors from the process of machine learning for pattern recognition or EEG signal analysis processing. Supervised learning based support vector machine or fuzzy c-mean (FCM) based on unsupervised learning have been used for classification in the field. However, conventional methods were hard to apply nonlinear dataset classification or required prior information for supervised learning. We solved this problems with proposed classification method using heuristic approach which took the minimal Euclidean distance between vectors, then we assumed them as same class and the others were another class. For the comparison, we used FCM, self-organizing map (SOM) based on artificial neural network (ANN). KEEL machine learning datset was used for simulation. We concluded that proposed method was superior than other algorithms.

The Optimization of Hybrid BCI Systems based on Blind Source Separation in Single Channel (단일 채널에서 블라인드 음원분리를 통한 하이브리드 BCI시스템 최적화)

  • Yang, Da-Lin;Nguyen, Trung-Hau;Kim, Jong-Jin;Chung, Wan-Young
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.19 no.1
    • /
    • pp.7-13
    • /
    • 2018
  • In the current study, we proposed an optimized brain-computer interface (BCI) which employed blind source separation (BBS) approach to remove noises. Thus motor imagery (MI) signal and steady state visual evoked potential (SSVEP) signal were easily to be detected due to enhancement in signal-to-noise ratio (SNR). Moreover, a combination between MI and SSVEP which is typically can increase the number of commands being generated in the current BCI. To reduce the computational time as well as to bring the BCI closer to real-world applications, the current system utilizes a single-channel EEG signal. In addition, a convolutional neural network (CNN) was used as the multi-class classification model. We evaluated the performance in term of accuracy between a non-BBS+BCI and BBS+BCI. Results show that the accuracy of the BBS+BCI is achieved $16.15{\pm}5.12%$ higher than that in the non-BBS+BCI by using BBS than non-used on. Overall, the proposed BCI system demonstrate a feasibility to be applied for multi-dimensional control applications with a comparable accuracy.

Development of Fire Detection Model for Underground Utility Facilities Using Deep Learning : Training Data Supplement and Bias Optimization (딥러닝 기반 지하공동구 화재 탐지 모델 개발 : 학습데이터 보강 및 편향 최적화)

  • Kim, Jeongsoo;Lee, Chan-Woo;Park, Seung-Hwa;Lee, Jong-Hyun;Hong, Chang-Hee
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.12
    • /
    • pp.320-330
    • /
    • 2020
  • Fire is difficult to achieve good performance in image detection using deep learning because of its high irregularity. In particular, there is little data on fire detection in underground utility facilities, which have poor light conditions and many objects similar to fire. These make fire detection challenging and cause low performance of deep learning models. Therefore, this study proposed a fire detection model using deep learning and estimated the performance of the model. The proposed model was designed using a combination of a basic convolutional neural network, Inception block of GoogleNet, and Skip connection of ResNet to optimize the deep learning model for fire detection under underground utility facilities. In addition, a training technique for the model was proposed. To examine the effectiveness of the method, the trained model was applied to fire images, which included fire and non-fire (which can be misunderstood as a fire) objects under the underground facilities or similar conditions, and results were analyzed. Metrics, such as precision and recall from deep learning models of other studies, were compared with those of the proposed model to estimate the model performance qualitatively. The results showed that the proposed model has high precision and recall for fire detection under low light intensity and both low erroneous and missing detection capabilities for things similar to fire.

Cloud Detection from Sentinel-2 Images Using DeepLabV3+ and Swin Transformer Models (DeepLabV3+와 Swin Transformer 모델을 이용한 Sentinel-2 영상의 구름탐지)

  • Kang, Jonggu;Park, Ganghyun;Kim, Geunah;Youn, Youjeong;Choi, Soyeon;Lee, Yangwon
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_2
    • /
    • pp.1743-1747
    • /
    • 2022
  • Sentinel-2 can be used as proxy data for the Korean Compact Advanced Satellite 500-4 (CAS500-4), also known as Agriculture and Forestry Satellite, in terms of spectral wavelengths and spatial resolution. This letter examined cloud detection for later use in the CAS500-4 based on deep learning technologies. DeepLabV3+, a traditional Convolutional Neural Network (CNN) model, and Shifted Windows (Swin) Transformer, a state-of-the-art (SOTA) Transformer model, were compared using 22,728 images provided by Radiant Earth Foundation (REF). Swin Transformer showed a better performance with a precision of 0.886 and a recall of 0.875, which is a balanced result, unbiased between over- and under-estimation. Deep learning-based cloud detection is expected to be a future operational module for CAS500-4 through optimization for the Korean Peninsula.

Prediction of the remaining time and time interval of pebbles in pebble bed HTGRs aided by CNN via DEM datasets

  • Mengqi Wu;Xu Liu;Nan Gui;Xingtuan Yang;Jiyuan Tu;Shengyao Jiang;Qian Zhao
    • Nuclear Engineering and Technology
    • /
    • v.55 no.1
    • /
    • pp.339-352
    • /
    • 2023
  • Prediction of the time-related traits of pebble flow inside pebble-bed HTGRs is of great significance for reactor operation and design. In this work, an image-driven approach with the aid of a convolutional neural network (CNN) is proposed to predict the remaining time of initially loaded pebbles and the time interval of paired flow images of the pebble bed. Two types of strategies are put forward: one is adding FC layers to the classic classification CNN models and using regression training, and the other is CNN-based deep expectation (DEX) by regarding the time prediction as a deep classification task followed by softmax expected value refinements. The current dataset is obtained from the discrete element method (DEM) simulations. Results show that the CNN-aided models generally make satisfactory predictions on the remaining time with the determination coefficient larger than 0.99. Among these models, the VGG19+DEX performs the best and its CumScore (proportion of test set with prediction error within 0.5s) can reach 0.939. Besides, the remaining time of additional test sets and new cases can also be well predicted, indicating good generalization ability of the model. In the task of predicting the time interval of image pairs, the VGG19+DEX model has also generated satisfactory results. Particularly, the trained model, with promising generalization ability, has demonstrated great potential in accurately and instantaneously predicting the traits of interest, without the need for additional computational intensive DEM simulations. Nevertheless, the issues of data diversity and model optimization need to be improved to achieve the full potential of the CNN-aided prediction tool.

Drape Simulation Estimation for Non-Linear Stiffness Model (비선형 강성 모델을 위한 드레이프 시뮬레이션 결과 추정)

  • Eungjune Shim;Eunjung Ju;Myung Geol Choi
    • Journal of the Korea Computer Graphics Society
    • /
    • v.29 no.3
    • /
    • pp.117-125
    • /
    • 2023
  • In the development of clothing design through virtual simulation, it is essential to minimize the differences between the virtual and the real world as much as possible. The most critical task to enhance the similarity between virtual and real garments is to find simulation parameters that can closely emulate the physical properties of the actual fabric in use. The simulation parameter optimization process requires manual tuning by experts, demanding high expertise and a significant amount of time. Especially, considerable time is consumed in repeatedly running simulations to check the results of applying the tuned simulation parameters. Recently, to tackle this issue, artificial neural network learning models have been proposed that swiftly estimate the results of drape test simulations, which are predominantly used for parameter tuning. In these earlier studies, relatively simple linear stiffness models were used, and instead of estimating the entirety of the drape mesh, they estimated only a portion of the mesh and interpolated the rest. However, there is still a scarcity of research on non-linear stiffness models, which are commonly used in actual garment design. In this paper, we propose a learning model for estimating the results of drape simulations for non-linear stiffness models. Our learning model estimates the full high-resolution mesh model of drape. To validate the performance of the proposed method, experiments were conducted using three different drape test methods, demonstrating high accuracy in estimation.

A study on end-to-end speaker diarization system using single-label classification (단일 레이블 분류를 이용한 종단 간 화자 분할 시스템 성능 향상에 관한 연구)

  • Jaehee Jung;Wooil Kim
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.6
    • /
    • pp.536-543
    • /
    • 2023
  • Speaker diarization, which labels for "who spoken when?" in speech with multiple speakers, has been studied on a deep neural network-based end-to-end method for labeling on speech overlap and optimization of speaker diarization models. Most deep neural network-based end-to-end speaker diarization systems perform multi-label classification problem that predicts the labels of all speakers spoken in each frame of speech. However, the performance of the multi-label-based model varies greatly depending on what the threshold is set to. In this paper, it is studied a speaker diarization system using single-label classification so that speaker diarization can be performed without thresholds. The proposed model estimate labels from the output of the model by converting speaker labels into a single label. To consider speaker label permutations in the training, the proposed model is used a combination of Permutation Invariant Training (PIT) loss and cross-entropy loss. In addition, how to add the residual connection structures to model is studied for effective learning of speaker diarization models with deep structures. The experiment used the Librispech database to generate and use simulated noise data for two speakers. When compared with the proposed method and baseline model using the Diarization Error Rate (DER) performance the proposed method can be labeling without threshold, and it has improved performance by about 20.7 %.

The Analysis and Design of Advanced Neurofuzzy Polynomial Networks (고급 뉴로퍼지 다항식 네트워크의 해석과 설계)

  • Park, Byeong-Jun;O, Seong-Gwon
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.39 no.3
    • /
    • pp.18-31
    • /
    • 2002
  • In this study, we introduce a concept of advanced neurofuzzy polynomial networks(ANFPN), a hybrid modeling architecture combining neurofuzzy networks(NFN) and polynomial neural networks(PNN). These networks are highly nonlinear rule-based models. The development of the ANFPN dwells on the technologies of Computational Intelligence(Cl), namely fuzzy sets, neural networks and genetic algorithms. NFN contributes to the formation of the premise part of the rule-based structure of the ANFPN. The consequence part of the ANFPN is designed using PNN. At the premise part of the ANFPN, NFN uses both the simplified fuzzy inference and error back-propagation learning rule. The parameters of the membership functions, learning rates and momentum coefficients are adjusted with the use of genetic optimization. As the consequence structure of ANFPN, PNN is a flexible network architecture whose structure(topology) is developed through learning. In particular, the number of layers and nodes of the PNN are not fixed in advance but is generated in a dynamic way. In this study, we introduce two kinds of ANFPN architectures, namely the basic and the modified one. Here the basic and the modified architecture depend on the number of input variables and the order of polynomial in each layer of PNN structure. Owing to the specific features of two combined architectures, it is possible to consider the nonlinear characteristics of process system and to obtain the better output performance with superb predictive ability. The availability and feasibility of the ANFPN are discussed and illustrated with the aid of two representative numerical examples. The results show that the proposed ANFPN can produce the model with higher accuracy and predictive ability than any other method presented previously.

Predicting Corporate Bankruptcy using Simulated Annealing-based Random Fores (시뮬레이티드 어니일링 기반의 랜덤 포레스트를 이용한 기업부도예측)

  • Park, Hoyeon;Kim, Kyoung-jae
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.155-170
    • /
    • 2018
  • Predicting a company's financial bankruptcy is traditionally one of the most crucial forecasting problems in business analytics. In previous studies, prediction models have been proposed by applying or combining statistical and machine learning-based techniques. In this paper, we propose a novel intelligent prediction model based on the simulated annealing which is one of the well-known optimization techniques. The simulated annealing is known to have comparable optimization performance to the genetic algorithms. Nevertheless, since there has been little research on the prediction and classification of business decision-making problems using the simulated annealing, it is meaningful to confirm the usefulness of the proposed model in business analytics. In this study, we use the combined model of simulated annealing and machine learning to select the input features of the bankruptcy prediction model. Typical types of combining optimization and machine learning techniques are feature selection, feature weighting, and instance selection. This study proposes a combining model for feature selection, which has been studied the most. In order to confirm the superiority of the proposed model in this study, we apply the real-world financial data of the Korean companies and analyze the results. The results show that the predictive accuracy of the proposed model is better than that of the naïve model. Notably, the performance is significantly improved as compared with the traditional decision tree, random forests, artificial neural network, SVM, and logistic regression analysis.

Cost Optimization of Doubly Reinforced Concrete Beam through Deep Reinforcement Learning without Labeled Data

  • Dongwoo Kim;Sangik Lee;Jonghyuk Lee;Byung-hun Seo;Dongsu Kim;Yejin Seo;Yerim Jo;Won Choi
    • International conference on construction engineering and project management
    • /
    • 2024.07a
    • /
    • pp.1322-1322
    • /
    • 2024
  • Reinforced concrete (RC) , a major contributor to resource depletion and harmful emissions, fuels research on optimizing its design. Optimizing RC structures is challenging due to the mix of discrete and continuous variables, hindering traditional differentiation-based methods. Thus, this study aims to optimize RC structures cost-effectively using deep reinforcement learning. When the Agent selects design variables, Environment checks design criteria based on KDS 14-20 code (South Korea) and calculates reward. The Agent updates its Neural Network with this reward. Target for optimization is a simply supported doubly RC beam, with design variables including cross-section dimensions, sizes and quantities of tension and compression reinforcement, and size of stirrups. We used 200,000 training sets and 336 test sets, each with live load, dead load, beam length variables. To exclude labeled data, multiple training iterations were conducted. In the initial training, the reward was the ratio of maximum possible cost at beam length to the designed structure's cost. Next iterations used the ratio of optimal values by the previous Agent to the current Agent as the reward. Training ended when the difference between optimal values from the previous and current Agent was within 1% for test data. Brute Force Algorithm was applied to the test set to calculate the actual cost-optimal design for validation. Results showed within 10% difference from actual optimal cost, indicating successful deep reinforcement learning application without labeled data. This study benefits the rapid and accurate calculation of optimized designs and construction processes in Building Information Modeling (BIM) applications.