• Title/Summary/Keyword: deep transfer learning

Search Result 252, Processing Time 0.024 seconds

Pedestrian Classification using CNN's Deep Features and Transfer Learning (CNN의 깊은 특징과 전이학습을 사용한 보행자 분류)

  • Chung, Soyoung;Chung, Min Gyo
    • Journal of Internet Computing and Services
    • /
    • v.20 no.4
    • /
    • pp.91-102
    • /
    • 2019
  • In autonomous driving systems, the ability to classify pedestrians in images captured by cameras is very important for pedestrian safety. In the past, after extracting features of pedestrians with HOG(Histogram of Oriented Gradients) or SIFT(Scale-Invariant Feature Transform), people classified them using SVM(Support Vector Machine). However, extracting pedestrian characteristics in such a handcrafted manner has many limitations. Therefore, this paper proposes a method to classify pedestrians reliably and effectively using CNN's(Convolutional Neural Network) deep features and transfer learning. We have experimented with both the fixed feature extractor and the fine-tuning methods, which are two representative transfer learning techniques. Particularly, in the fine-tuning method, we have added a new scheme, called M-Fine(Modified Fine-tuning), which divideslayers into transferred parts and non-transferred parts in three different sizes, and adjusts weights only for layers belonging to non-transferred parts. Experiments on INRIA Person data set with five CNN models(VGGNet, DenseNet, Inception V3, Xception, and MobileNet) showed that CNN's deep features perform better than handcrafted features such as HOG and SIFT, and that the accuracy of Xception (threshold = 0.5) isthe highest at 99.61%. MobileNet, which achieved similar performance to Xception and learned 80% fewer parameters, was the best in terms of efficiency. Among the three transfer learning schemes tested above, the performance of the fine-tuning method was the best. The performance of the M-Fine method was comparable to or slightly lower than that of the fine-tuningmethod, but higher than that of the fixed feature extractor method.

CNN model transition learning comparative analysis based on deep learning for image classification (이미지 분류를 위한 딥러닝 기반 CNN모델 전이 학습 비교 분석)

  • Lee, Dong-jun;Jeon, Seung-Je;Lee, DongHwi
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.370-373
    • /
    • 2022
  • Recently, various deep learning framework models such as Tensorflow, Pytorch, Keras, etc. have appeared. In addition, CNN (Convolutional Neural Network) is applied to image recognition using frameworks such as Tensorflow, Pytorch, and Keras, and the optimization model in image classification is mainly used. In this paper, based on the results of training the CNN model with the Paitotchi and tensor flow frameworks most often used in the field of deep learning image recognition, the two frameworks are compared and analyzed for image analysis. Derived an optimized framework.

  • PDF

Predicting Dynamic Response of a Railway Bridge Using Transfer-Learning Technique (전이학습 기법을 이용한 철도교량의 동적응답 예측)

  • Minsu Kim;Sanghyun Choi
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.36 no.1
    • /
    • pp.39-48
    • /
    • 2023
  • Because a railway bridge is designed over a long period of time and covers a large site, it involves various environmental factors and uncertainties. For this reason, design changes often occur, even if the design was thoroughly reviewed in the initial design stage. In particular, design changes of large-scale facilities, such as railway bridges, consume significant time and cost, and it is extremely inefficient to repeat all the procedures each time. In this study, a technique that can improve the efficiency of learning after design change was developed by utilizing the learning result before design change through transfer learning among deep-learning algorithms. For analysis, scenarios were created, and a database was built using a previously developed railway bridge deep-learning-based prediction system. The proposed method results in similar accuracy when learning only 1000 data points in the new domain compared with the 8000 data points used for learning in the old domain before the design change. Moreover, it was confirmed that it has a faster convergence speed.

MULTI-APERTURE IMAGE PROCESSING USING DEEP LEARNING

  • GEONHO HWANG;CHANG HOON SONG;TAE KYUNG LEE;HOJUN NA;MYUNGJOO KANG
    • Journal of the Korean Society for Industrial and Applied Mathematics
    • /
    • v.27 no.1
    • /
    • pp.56-74
    • /
    • 2023
  • In order to obtain practical and high-quality satellite images containing high-frequency components, a large aperture optical system is required, which has a limitation in that it greatly increases the payload weight. As an attempt to overcome the problem, many multi-aperture optical systems have been proposed, but in many cases, these optical systems do not include high-frequency components in all directions, and making such an high-quality image is an ill-posed problem. In this paper, we use deep learning to overcome the limitation. A deep learning model receives low-quality images as input, estimates the Point Spread Function, PSF, and combines them to output a single high-quality image. We model images obtained from three rectangular apertures arranged in a regular polygon shape. We also propose the Modulation Transfer Function Loss, MTF Loss, which can capture the high-frequency components of the images. We present qualitative and quantitative results obtained through experiments.

A Study on Transferring Cloud Dataset for Smoke Extraction Based on Deep Learning (딥러닝 기반 연기추출을 위한 구름 데이터셋의 전이학습에 대한 연구)

  • Kim, Jiyong;Kwak, Taehong;Kim, Yongil
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_2
    • /
    • pp.695-706
    • /
    • 2022
  • Medium and high-resolution optical satellites have proven their effectiveness in detecting wildfire areas. However, smoke plumes generated by wildfire scatter visible light incidents on the surface, thereby interrupting accurate monitoring of the area where wildfire occurs. Therefore, a technology to extract smoke in advance is required. Deep learning technology is expected to improve the accuracy of smoke extraction, but the lack of training datasets limits the application. However, for clouds, which have a similar property of scattering visible light, a large amount of training datasets has been accumulated. The purpose of this study is to develop a smoke extraction technique using deep learning, and the limits due to the lack of datasets were overcome by using a cloud dataset on transfer learning. To check the effectiveness of transfer learning, a small-scale smoke extraction training set was made, and the smoke extraction performance was compared before and after applying transfer learning using a public cloud dataset. As a result, not only the performance in the visible light wavelength band was enhanced but also in the near infrared (NIR) and short-wave infrared (SWIR). Through the results of this study, it is expected that the lack of datasets, which is a critical limit for using deep learning on smoke extraction, can be solved, and therefore, through the advancement of smoke extraction technology, it will be possible to present an advantage in monitoring wildfires.

Data Augmentation Techniques of Power Facilities for Improve Deep Learning Performance

  • Jang, Seungmin;Son, Seungwoo;Kim, Bongsuck
    • KEPCO Journal on Electric Power and Energy
    • /
    • v.7 no.2
    • /
    • pp.323-328
    • /
    • 2021
  • Diagnostic models are required. Data augmentation is one of the best ways to improve deep learning performance. Traditional augmentation techniques that modify image brightness or spatial information are difficult to achieve great results. To overcome this, a generative adversarial network (GAN) technology that generates virtual data to increase deep learning performance has emerged. GAN can create realistic-looking fake images by competitive learning two networks, a generator that creates fakes and a discriminator that determines whether images are real or fake made by the generator. GAN is being used in computer vision, IT solutions, and medical imaging fields. It is essential to secure additional learning data to advance deep learning-based fault diagnosis solutions in the power industry where facilities are strictly maintained more than other industries. In this paper, we propose a method for generating power facility images using GAN and a strategy for improving performance when only used a small amount of data. Finally, we analyze the performance of the augmented image to see if it could be utilized for the deep learning-based diagnosis system or not.

A Study on Realtime Drone Object Detection Using On-board Deep Learning (온-보드에서의 딥러닝을 활용한 드론의 실시간 객체 인식 연구)

  • Lee, Jang-Woo;Kim, Joo-Young;Kim, Jae-Kyung;Kwon, Cheol-Hee
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.49 no.10
    • /
    • pp.883-892
    • /
    • 2021
  • This paper provides a process for developing deep learning-based aerial object detection models that can run in realtime on onboard. To improve object detection performance, we pre-process and augment the training data in the training stage. In addition, we perform transfer learning and apply a weighted cross-entropy method to reduce the variations of detection performance for each class. To improve the inference speed, we have generated inference acceleration engines with quantization. Then, we analyze the real-time performance and detection performance on custom aerial image dataset to verify generalization.

Transfer Learning Using Convolutional Neural Network Architectures for Glioma Classification from MRI Images

  • Kulkarni, Sunita M.;Sundari, G.
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.2
    • /
    • pp.198-204
    • /
    • 2021
  • Glioma is one of the common types of brain tumors starting in the brain's glial cell. These tumors are classified into low-grade or high-grade tumors. Physicians analyze the stages of brain tumors and suggest treatment to the patient. The status of the tumor has an importance in the treatment. Nowadays, computerized systems are used to analyze and classify brain tumors. The accurate grading of the tumor makes sense in the treatment of brain tumors. This paper aims to develop a classification of low-grade glioma and high-grade glioma using a deep learning algorithm. This system utilizes four transfer learning algorithms, i.e., AlexNet, GoogLeNet, ResNet18, and ResNet50, for classification purposes. Among these algorithms, ResNet18 shows the highest classification accuracy of 97.19%.

K-Means Clustering with Deep Learning for Fingerprint Class Type Prediction

  • Mukoya, Esther;Rimiru, Richard;Kimwele, Michael;Mashava, Destine
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.3
    • /
    • pp.29-36
    • /
    • 2022
  • In deep learning classification tasks, most models frequently assume that all labels are available for the training datasets. As such strategies to learn new concepts from unlabeled datasets are scarce. In fingerprint classification tasks, most of the fingerprint datasets are labelled using the subject/individual and fingerprint datasets labelled with finger type classes are scarce. In this paper, authors have developed approaches of classifying fingerprint images using the majorly known fingerprint classes. Our study provides a flexible method to learn new classes of fingerprints. Our classifier model combines both the clustering technique and use of deep learning to cluster and hence label the fingerprint images into appropriate classes. The K means clustering strategy explores the label uncertainty and high-density regions from unlabeled data to be clustered. Using similarity index, five clusters are created. Deep learning is then used to train a model using a publicly known fingerprint dataset with known finger class types. A prediction technique is then employed to predict the classes of the clusters from the trained model. Our proposed model is better and has less computational costs in learning new classes and hence significantly saving on labelling costs of fingerprint images.

An Optimized Deep Learning Techniques for Analyzing Mammograms

  • Satish Babu Bandaru;Natarajasivan. D;Rama Mohan Babu. G
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.7
    • /
    • pp.39-48
    • /
    • 2023
  • Breast cancer screening makes extensive utilization of mammography. Even so, there has been a lot of debate with regards to this application's starting age as well as screening interval. The deep learning technique of transfer learning is employed for transferring the knowledge learnt from the source tasks to the target tasks. For the resolution of real-world problems, deep neural networks have demonstrated superior performance in comparison with the standard machine learning algorithms. The architecture of the deep neural networks has to be defined by taking into account the problem domain knowledge. Normally, this technique will consume a lot of time as well as computational resources. This work evaluated the efficacy of the deep learning neural network like Visual Geometry Group Network (VGG Net) Residual Network (Res Net), as well as inception network for classifying the mammograms. This work proposed optimization of ResNet with Teaching Learning Based Optimization (TLBO) algorithm's in order to predict breast cancers by means of mammogram images. The proposed TLBO-ResNet, an optimized ResNet with faster convergence ability when compared with other evolutionary methods for mammogram classification.