• Title/Summary/Keyword: deep neural net

Search Result 327, Processing Time 0.078 seconds

A Study on the Classification of Surface Defect Based on Deep Convolution Network and Transfer-learning (신경망과 전이학습 기반 표면 결함 분류에 관한 연구)

  • Kim, Sung Joo;Kim, Gyung Bum
    • Journal of the Semiconductor & Display Technology
    • /
    • v.20 no.1
    • /
    • pp.64-69
    • /
    • 2021
  • In this paper, a method for improving the defect classification performance in low contrast, ununiformity and featureless steel plate surfaces has been studied based on deep convolution neural network and transfer-learning neural network. The steel plate surface images have low contrast, ununiformity, and featureless, so that the contrast between defect and defect-free regions are not discriminated. These characteristics make it difficult to extract the feature of the surface defect image. A classifier based on a deep convolution neural network is constructed to extract features automatically for effective classification of images with these characteristics. As results of the experiment, AlexNet-based transfer-learning classifier showed excellent classification performance of 99.43% with less than 160 seconds of training time. The proposed classification system showed excellent classification performance for low contrast, ununiformity, and featureless surface images.

Very short-term rainfall prediction based on radar image learning using deep neural network (심층신경망을 이용한 레이더 영상 학습 기반 초단시간 강우예측)

  • Yoon, Seongsim;Park, Heeseong;Shin, Hongjoon
    • Journal of Korea Water Resources Association
    • /
    • v.53 no.12
    • /
    • pp.1159-1172
    • /
    • 2020
  • This study applied deep convolution neural network based on U-Net and SegNet using long period weather radar data to very short-term rainfall prediction. And the results were compared and evaluated with the translation model. For training and validation of deep neural network, Mt. Gwanak and Mt. Gwangdeoksan radar data were collected from 2010 to 2016 and converted to a gray-scale image file in an HDF5 format with a 1km spatial resolution. The deep neural network model was trained to predict precipitation after 10 minutes by using the four consecutive radar image data, and the recursive method of repeating forecasts was applied to carry out lead time 60 minutes with the pretrained deep neural network model. To evaluate the performance of deep neural network prediction model, 24 rain cases in 2017 were forecast for rainfall up to 60 minutes in advance. As a result of evaluating the predicted performance by calculating the mean absolute error (MAE) and critical success index (CSI) at the threshold of 0.1, 1, and 5 mm/hr, the deep neural network model showed better performance in the case of rainfall threshold of 0.1, 1 mm/hr in terms of MAE, and showed better performance than the translation model for lead time 50 minutes in terms of CSI. In particular, although the deep neural network prediction model performed generally better than the translation model for weak rainfall of 5 mm/hr or less, the deep neural network prediction model had limitations in predicting distinct precipitation characteristics of high intensity as a result of the evaluation of threshold of 5 mm/hr. The longer lead time, the spatial smoothness increase with lead time thereby reducing the accuracy of rainfall prediction The translation model turned out to be superior in predicting the exceedance of higher intensity thresholds (> 5 mm/hr) because it preserves distinct precipitation characteristics, but the rainfall position tends to shift incorrectly. This study are expected to be helpful for the improvement of radar rainfall prediction model using deep neural networks in the future. In addition, the massive weather radar data established in this study will be provided through open repositories for future use in subsequent studies.

Real-world noisy image denoising using deep residual U-Net structure (깊은 잔차 U-Net 구조를 이용한 실제 카메라 잡음 영상 디노이징)

  • Jang, Yeongil;Cho, Nam Ik
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2019.11a
    • /
    • pp.119-121
    • /
    • 2019
  • 부가적 백색 잡음 모델(additive white Gaussian noise, AWGN에서 학습된 깊은 신경만 (deep neural networks)을 이용한 잡음 제거기는 제거하려는 잡음이 AWGN인 경우에는 뛰어난 성능을 보이지만 실제 카메라 잡음에 대해서 잡음 제거를 시도하였을 때는 성능이 크게 저하된다. 본 논문은 U-Net 구조의 깊은 인공신경망 모델에 residual block을 결합함으로서 실제 카메라 영상에서 기존 알고리즘보다 뛰어난 성능을 지니는 신경망을 제안하다. 제안한 방법을 통해 Darmstadt Noise Dataset에서 PSNR과 SSIM 모두 CBDNet 대비 향상됨을 확인하였다.

  • PDF

Estimation of residual stress in dissimilar metals welding using deep fuzzy neural networks with rule-dropout

  • Ji Hun Park;Man Gyun Na
    • Nuclear Engineering and Technology
    • /
    • v.56 no.10
    • /
    • pp.4149-4157
    • /
    • 2024
  • Welding processes are used to connect several components in nuclear power plants. These welding processes can induce residual stress in welding joints, which has been identified as a significant factor in primary water stress corrosion cracking. Consequently, the assessment of welding residual stress plays a crucial role in determining the structural integrity of welded joints. In this study, a deep fuzzy neural networks (DFNN) with a rule-dropout method, which is an artificial intelligence (AI) method, was used to predict the residual stress of dissimilar metal welding. ABAQUS, a finite element analysis program, was used as the data collection tool to develop the AI model, and 6300 data instances were collected under 150 analysis conditions. A rule-dropout method and genetic algorithm were used to optimize the estimation performance of the DFNN model. DFNN with the rule-dropout model was compared to a deep neural network method, known as a general deep learning method, to evaluate the estimation performance of DFNN. In addition, a fuzzy neural network method and a cascaded support vector regression method conducted in previous studies were compared. Consequently, the estimation performance of the DFNN with the rule-dropout model was better than those of the comparison methods. The welding residual stress estimation results of this study are expected to contribute to the evaluation of the structural integrity of welded joints.

A Manually Captured and Modified Phone Screen Image Dataset for Widget Classification on CNNs

  • Byun, SungChul;Han, Seong-Soo;Jeong, Chang-Sung
    • Journal of Information Processing Systems
    • /
    • v.18 no.2
    • /
    • pp.197-207
    • /
    • 2022
  • The applications and user interfaces (UIs) of smart mobile devices are constantly diversifying. For example, deep learning can be an innovative solution to classify widgets in screen images for increasing convenience. To this end, the present research leverages captured images and the ReDraw dataset to write deep learning datasets for image classification purposes. First, as the validation for datasets using ResNet50 and EfficientNet, the experiments show that the dataset composed in this study is helpful for classification according to a widget's functionality. An implementation for widget detection and classification on RetinaNet and EfficientNet is then executed. Finally, the research suggests the Widg-C and Widg-D datasets-a deep learning dataset for identifying the widgets of smart devices-and implementing them for use with representative convolutional neural network models.

Text Classification on Social Network Platforms Based on Deep Learning Models

  • YA, Chen;Tan, Juan;Hoekyung, Jung
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.1
    • /
    • pp.9-16
    • /
    • 2023
  • The natural language on social network platforms has a certain front-to-back dependency in structure, and the direct conversion of Chinese text into a vector makes the dimensionality very high, thereby resulting in the low accuracy of existing text classification methods. To this end, this study establishes a deep learning model that combines a big data ultra-deep convolutional neural network (UDCNN) and long short-term memory network (LSTM). The deep structure of UDCNN is used to extract the features of text vector classification. The LSTM stores historical information to extract the context dependency of long texts, and word embedding is introduced to convert the text into low-dimensional vectors. Experiments are conducted on the social network platforms Sogou corpus and the University HowNet Chinese corpus. The research results show that compared with CNN + rand, LSTM, and other models, the neural network deep learning hybrid model can effectively improve the accuracy of text classification.

Study on the Surface Defect Classification of Al 6061 Extruded Material By Using CNN-Based Algorithms (CNN을 이용한 Al 6061 압출재의 표면 결함 분류 연구)

  • Kim, S.B.;Lee, K.A.
    • Transactions of Materials Processing
    • /
    • v.31 no.4
    • /
    • pp.229-239
    • /
    • 2022
  • Convolution Neural Network(CNN) is a class of deep learning algorithms and can be used for image analysis. In particular, it has excellent performance in finding the pattern of images. Therefore, CNN is commonly applied for recognizing, learning and classifying images. In this study, the surface defect classification performance of Al 6061 extruded material using CNN-based algorithms were compared and evaluated. First, the data collection criteria were suggested and a total of 2,024 datasets were prepared. And they were randomly classified into 1,417 learning data and 607 evaluation data. After that, the size and quality of the training data set were improved using data augmentation techniques to increase the performance of deep learning. The CNN-based algorithms used in this study were VGGNet-16, VGGNet-19, ResNet-50 and DenseNet-121. The evaluation of the defect classification performance was made by comparing the accuracy, loss, and learning speed using verification data. The DenseNet-121 algorithm showed better performance than other algorithms with an accuracy of 99.13% and a loss value of 0.037. This was due to the structural characteristics of the DenseNet model, and the information loss was reduced by acquiring information from all previous layers for image identification in this algorithm. Based on the above results, the possibility of machine vision application of CNN-based model for the surface defect classification of Al extruded materials was also discussed.

A Remote Sensing Scene Classification Model Based on EfficientNetV2L Deep Neural Networks

  • Aljabri, Atif A.;Alshanqiti, Abdullah;Alkhodre, Ahmad B.;Alzahem, Ayyub;Hagag, Ahmed
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.10
    • /
    • pp.406-412
    • /
    • 2022
  • Scene classification of very high-resolution (VHR) imagery can attribute semantics to land cover in a variety of domains. Real-world application requirements have not been addressed by conventional techniques for remote sensing image classification. Recent research has demonstrated that deep convolutional neural networks (CNNs) are effective at extracting features due to their strong feature extraction capabilities. In order to improve classification performance, these approaches rely primarily on semantic information. Since the abstract and global semantic information makes it difficult for the network to correctly classify scene images with similar structures and high interclass similarity, it achieves a low classification accuracy. We propose a VHR remote sensing image classification model that uses extracts the global feature from the original VHR image using an EfficientNet-V2L CNN pre-trained to detect similar classes. The image is then classified using a multilayer perceptron (MLP). This method was evaluated using two benchmark remote sensing datasets: the 21-class UC Merced, and the 38-class PatternNet. As compared to other state-of-the-art models, the proposed model significantly improves performance.

ResNet-Based Simulations for a Heat-Transfer Model Involving an Imperfect Contact

  • Guangxing, Wang;Gwanghyun, Jo;Seong-Yoon, Shin
    • Journal of information and communication convergence engineering
    • /
    • v.20 no.4
    • /
    • pp.303-308
    • /
    • 2022
  • Simulating the heat transfer in a composite material is an important topic in material science. Difficulties arise from the fact that adjacent materials cannot match perfectly, resulting in discontinuity in the temperature variables. Although there have been several numerical methods for solving the heat-transfer problem in imperfect contact conditions, the methods known so far are complicated to implement, and the computational times are non-negligible. In this study, we developed a ResNet-type deep neural network for simulating a heat transfer model in a composite material. To train the neural network, we generated datasets by numerically solving the heat-transfer equations with Kapitza thermal resistance conditions. Because datasets involve various configurations of composite materials, our neural networks are robust to the shapes of material-material interfaces. Our algorithm can predict the thermal behavior in real time once the networks are trained. The performance of the proposed neural networks is documented, where the root mean square error (RMSE) and mean absolute error (MAE) are below 2.47E-6, and 7.00E-4, respectively.

Crack segmentation in high-resolution images using cascaded deep convolutional neural networks and Bayesian data fusion

  • Tang, Wen;Wu, Rih-Teng;Jahanshahi, Mohammad R.
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.221-235
    • /
    • 2022
  • Manual inspection of steel box girders on long span bridges is time-consuming and labor-intensive. The quality of inspection relies on the subjective judgements of the inspectors. This study proposes an automated approach to detect and segment cracks in high-resolution images. An end-to-end cascaded framework is proposed to first detect the existence of cracks using a deep convolutional neural network (CNN) and then segment the crack using a modified U-Net encoder-decoder architecture. A Naïve Bayes data fusion scheme is proposed to reduce the false positives and false negatives effectively. To generate the binary crack mask, first, the original images are divided into 448 × 448 overlapping image patches where these image patches are classified as cracks versus non-cracks using a deep CNN. Next, a modified U-Net is trained from scratch using only the crack patches for segmentation. A customized loss function that consists of binary cross entropy loss and the Dice loss is introduced to enhance the segmentation performance. Additionally, a Naïve Bayes fusion strategy is employed to integrate the crack score maps from different overlapping crack patches and to decide whether a pixel is crack or not. Comprehensive experiments have demonstrated that the proposed approach achieves an 81.71% mean intersection over union (mIoU) score across 5 different training/test splits, which is 7.29% higher than the baseline reference implemented with the original U-Net.