• Title/Summary/Keyword: deep neural net

Search Result 327, Processing Time 0.024 seconds

Compressed-Sensing Cardiac CINE MRI using Neural Network with Transfer Learning (전이학습을 수행한 신경망을 사용한 압축센싱 심장 자기공명영상)

  • Park, Seong-Jae;Yoon, Jong-Hyun;Ahn, Chang-Beom
    • Journal of IKEEE
    • /
    • v.23 no.4
    • /
    • pp.1408-1414
    • /
    • 2019
  • Deep artificial neural network with transfer learning is applied to compressed sensing cardiovascular MRI. Transfer learning is a method that utilizes structure, filter kernels, and weights of the network used in prior learning for current learning or application. The transfer learning is useful in accelerating learning speed, and in generalization of the neural network when learning data is limited. From a cardiac MRI experiment, with 8 healthy volunteers, the neural network with transfer learning was able to reduce learning time by a factor of more than five compared to that with standalone learning. Using test data set, reconstructed images with transfer learning showed lower normalized mean square error and better image quality compared to those without transfer learning.

U-net and Residual-based Cycle-GAN for Improving Object Transfiguration Performance (물체 변형 성능을 향상하기 위한 U-net 및 Residual 기반의 Cycle-GAN)

  • Kim, Sewoon;Park, Kwang-Hyun
    • The Journal of Korea Robotics Society
    • /
    • v.13 no.1
    • /
    • pp.1-7
    • /
    • 2018
  • The image-to-image translation is one of the deep learning applications using image data. In this paper, we aim at improving the performance of object transfiguration which transforms a specific object in an image into another specific object. For object transfiguration, it is required to transform only the target object and maintain background images. In the existing results, however, it is observed that other parts in the image are also transformed. In this paper, we have focused on the structure of artificial neural networks that are frequently used in the existing methods and have improved the performance by adding constraints to the exiting structure. We also propose the advanced structure that combines the existing structures to maintain their advantages and complement their drawbacks. The effectiveness of the proposed methods are shown in experimental results.

ADD-Net: Attention Based 3D Dense Network for Action Recognition

  • Man, Qiaoyue;Cho, Young Im
    • Journal of the Korea Society of Computer and Information
    • /
    • v.24 no.6
    • /
    • pp.21-28
    • /
    • 2019
  • Recent years with the development of artificial intelligence and the success of the deep model, they have been deployed in all fields of computer vision. Action recognition, as an important branch of human perception and computer vision system research, has attracted more and more attention. Action recognition is a challenging task due to the special complexity of human movement, the same movement may exist between multiple individuals. The human action exists as a continuous image frame in the video, so action recognition requires more computational power than processing static images. And the simple use of the CNN network cannot achieve the desired results. Recently, the attention model has achieved good results in computer vision and natural language processing. In particular, for video action classification, after adding the attention model, it is more effective to focus on motion features and improve performance. It intuitively explains which part the model attends to when making a particular decision, which is very helpful in real applications. In this paper, we proposed a 3D dense convolutional network based on attention mechanism(ADD-Net), recognition of human motion behavior in the video.

Scaling Up Face Masks Classification Using a Deep Neural Network and Classical Method Inspired Hybrid Technique

  • Kumar, Akhil;Kalia, Arvind;Verma, Kinshuk;Sharma, Akashdeep;Kaushal, Manisha;Kalia, Aayushi
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.11
    • /
    • pp.3658-3679
    • /
    • 2022
  • Classification of persons wearing and not wearing face masks in images has emerged as a new computer vision problem during the COVID-19 pandemic. In order to address this problem and scale up the research in this domain, in this paper a hybrid technique by employing ResNet-101 and multi-layer perceptron (MLP) classifier has been proposed. The proposed technique is tested and validated on a self-created face masks classification dataset and a standard dataset. On self-created dataset, the proposed technique achieved a classification accuracy of 97.3%. To embrace the proposed technique, six other state-of-the-art CNN feature extractors with six other classical machine learning classifiers have been tested and compared with the proposed technique. The proposed technique achieved better classification accuracy and 1-6% higher precision, recall, and F1 score as compared to other tested deep feature extractors and machine learning classifiers.

Feasibility study of deep learning based radiosensitivity prediction model of National Cancer Institute-60 cell lines using gene expression

  • Kim, Euidam;Chung, Yoonsun
    • Nuclear Engineering and Technology
    • /
    • v.54 no.4
    • /
    • pp.1439-1448
    • /
    • 2022
  • Background: We investigated the feasibility of in vitro radiosensitivity prediction with gene expression using deep learning. Methods: A microarray gene expression of the National Cancer Institute-60 (NCI-60) panel was acquired from the Gene Expression Omnibus. The clonogenic surviving fractions at an absorbed dose of 2 Gy (SF2) from previous publications were used to measure in vitro radiosensitivity. The radiosensitivity prediction model was based on the convolutional neural network. The 6-fold cross-validation (CV) was applied to train and validate the model. Then, the leave-one-out cross-validation (LOOCV) was applied by using the large-errored samples as a validation set, to determine whether the error was from the high bias of the folded CV. The criteria for correct prediction were defined as an absolute error<0.01 or a relative error<10%. Results: Of the 174 triplicated samples of NCI-60, 171 samples were correctly predicted with the folded CV. Through an additional LOOCV, one more sample was correctly predicted, representing a prediction accuracy of 98.85% (172 out of 174 samples). The average relative error and absolute errors of 172 correctly predicted samples were 1.351±1.875% and 0.00596±0.00638, respectively. Conclusion: We demonstrated the feasibility of a deep learning-based in vitro radiosensitivity prediction using gene expression.

Toward Practical Augmentation of Raman Spectra for Deep Learning Classification of Contamination in HDD

  • Seksan Laitrakun;Somrudee Deepaisarn;Sarun Gulyanon;Chayud Srisumarnk;Nattapol Chiewnawintawat;Angkoon Angkoonsawaengsuk;Pakorn Opaprakasit;Jirawan Jindakaew;Narisara Jaikaew
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.3
    • /
    • pp.208-215
    • /
    • 2023
  • Deep learning techniques provide powerful solutions to several pattern-recognition problems, including Raman spectral classification. However, these networks require large amounts of labeled data to perform well. Labeled data, which are typically obtained in a laboratory, can potentially be alleviated by data augmentation. This study investigated various data augmentation techniques and applied multiple deep learning methods to Raman spectral classification. Raman spectra yield fingerprint-like information about chemical compositions, but are prone to noise when the particles of the material are small. Five augmentation models were investigated to build robust deep learning classifiers: weighted sums of spectral signals, imitated chemical backgrounds, extended multiplicative signal augmentation, and generated Gaussian and Poisson-distributed noise. We compared the performance of nine state-of-the-art convolutional neural networks with all the augmentation techniques. The LeNet5 models with background noise augmentation yielded the highest accuracy when tested on real-world Raman spectral classification at 88.33% accuracy. A class activation map of the model was generated to provide a qualitative observation of the results.

Deep-learning-based system-scale diagnosis of a nuclear power plant with multiple infrared cameras

  • Ik Jae Jin;Do Yeong Lim;In Cheol Bang
    • Nuclear Engineering and Technology
    • /
    • v.55 no.2
    • /
    • pp.493-505
    • /
    • 2023
  • Comprehensive condition monitoring of large industry systems such as nuclear power plants (NPPs) is essential for safety and maintenance. In this study, we developed novel system-scale diagnostic technology based on deep-learning and IR thermography that can efficiently and cost-effectively classify system conditions using compact Raspberry Pi and IR sensors. This diagnostic technology can identify the presence of an abnormality or accident in whole system, and when an accident occurs, the type of accident and the location of the abnormality can be identified in real-time. For technology development, the experiment for the thermal image measurement and performance validation of major components at each accident condition of NPPs was conducted using a thermal-hydraulic integral effect test facility with compact infrared sensor modules. These thermal images were used for training of deep-learning model, convolutional neural networks (CNN), which is effective for image processing. As a result, a proposed novel diagnostic was developed that can perform diagnosis of components, whole system and accident classification using thermal images. The optimal model was derived based on the modern CNN model and performed prompt and accurate condition monitoring of component and whole system diagnosis, and accident classification. This diagnostic technology is expected to be applied to comprehensive condition monitoring of nuclear power plants for safety.

Design of Speech Enhancement U-Net for Embedded Computing (임베디드 연산을 위한 잡음에서 음성추출 U-Net 설계)

  • Kim, Hyun-Don
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.15 no.5
    • /
    • pp.227-234
    • /
    • 2020
  • In this paper, we propose wav-U-Net to improve speech enhancement in heavy noisy environments, and it has implemented three principal techniques. First, as input data, we use 128 modified Mel-scale filter banks which can reduce computational burden instead of 512 frequency bins. Mel-scale aims to mimic the non-linear human ear perception of sound by being more discriminative at lower frequencies and less discriminative at higher frequencies. Therefore, Mel-scale is the suitable feature considering both performance and computing power because our proposed network focuses on speech signals. Second, we add a simple ResNet as pre-processing that helps our proposed network make estimated speech signals clear and suppress high-frequency noises. Finally, the proposed U-Net model shows significant performance regardless of the kinds of noise. Especially, despite using a single channel, we confirmed that it can well deal with non-stationary noises whose frequency properties are dynamically changed, and it is possible to estimate speech signals from noisy speech signals even in extremely noisy environments where noises are much lauder than speech (less than SNR 0dB). The performance on our proposed wav-U-Net was improved by about 200% on SDR and 460% on NSDR compared to the conventional Jansson's wav-U-Net. Also, it was confirmed that the processing time of out wav-U-Net with 128 modified Mel-scale filter banks was about 2.7 times faster than the common wav-U-Net with 512 frequency bins as input values.

A Video Smoke Detection Algorithm Based on Cascade Classification and Deep Learning

  • Nguyen, Manh Dung;Kim, Dongkeun;Ro, Soonghwan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.12
    • /
    • pp.6018-6033
    • /
    • 2018
  • Fires are a common cause of catastrophic personal injuries and devastating property damage. Every year, many fires occur and threaten human lives and property around the world. Providing early important sign for early fire detection, and therefore the detection of smoke is always the first step in fire-alarm systems. In this paper we propose an automatic smoke detection system built on camera surveillance and image processing technologies. The key features used in our algorithm are to detect and track smoke as moving objects and distinguish smoke from non-smoke objects using a convolutional neural network (CNN) model for cascade classification. The results of our experiment, in comparison with those of some earlier studies, show that the proposed algorithm is very effective not only in detecting smoke, but also in reducing false positives.

Refinement of Ground Truth Data for X-ray Coronary Artery Angiography (CAG) using Active Contour Model

  • Dongjin Han;Youngjoon Park
    • International journal of advanced smart convergence
    • /
    • v.12 no.4
    • /
    • pp.134-141
    • /
    • 2023
  • We present a novel method aimed at refining ground truth data through regularization and modification, particularly applicable when working with the original ground truth set. Enhancing the performance of deep neural networks is achieved by applying regularization techniques to the existing ground truth data. In many machine learning tasks requiring pixel-level segmentation sets, accurately delineating objects is vital. However, it proves challenging for thin and elongated objects such as blood vessels in X-ray coronary angiography, often resulting in inconsistent generation of ground truth data. This method involves an analysis of the quality of training set pairs - comprising images and ground truth data - to automatically regulate and modify the boundaries of ground truth segmentation. Employing the active contour model and a recursive ground truth generation approach results in stable and precisely defined boundary contours. Following the regularization and adjustment of the ground truth set, there is a substantial improvement in the performance of deep neural networks.