• 제목/요약/키워드: VGG16

검색결과 122건 처리시간 0.028초

Comparison of Fine-Tuned Convolutional Neural Networks for Clipart Style Classification

  • Lee, Seungbin;Kim, Hyungon;Seok, Hyekyoung;Nang, Jongho
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제9권4호
    • /
    • pp.1-7
    • /
    • 2017
  • Clipart is artificial visual contents that are created using various tools such as Illustrator to highlight some information. Here, the style of the clipart plays a critical role in determining how it looks. However, previous studies on clipart are focused only on the object recognition [16], segmentation, and retrieval of clipart images using hand-craft image features. Recently, some clipart classification researches based on the style similarity using CNN have been proposed, however, they have used different CNN-models and experimented with different benchmark dataset so that it is very hard to compare their performances. This paper presents an experimental analysis of the clipart classification based on the style similarity with two well-known CNN-models (Inception Resnet V2 [13] and VGG-16 [14] and transfers learning with the same benchmark dataset (Microsoft Style Dataset 3.6K). From this experiment, we find out that the accuracy of Inception Resnet V2 is better than VGG for clipart style classification because of its deep nature and convolution map with various sizes in parallel. We also find out that the end-to-end training can improve the accuracy more than 20% in both CNN models.

Waste Classification by Fine-Tuning Pre-trained CNN and GAN

  • Alsabei, Amani;Alsayed, Ashwaq;Alzahrani, Manar;Al-Shareef, Sarah
    • International Journal of Computer Science & Network Security
    • /
    • 제21권8호
    • /
    • pp.65-70
    • /
    • 2021
  • Waste accumulation is becoming a significant challenge in most urban areas and if it continues unchecked, is poised to have severe repercussions on our environment and health. The massive industrialisation in our cities has been followed by a commensurate waste creation that has become a bottleneck for even waste management systems. While recycling is a viable solution for waste management, it can be daunting to classify waste material for recycling accurately. In this study, transfer learning models were proposed to automatically classify wastes based on six materials (cardboard, glass, metal, paper, plastic, and trash). The tested pre-trained models were ResNet50, VGG16, InceptionV3, and Xception. Data augmentation was done using a Generative Adversarial Network (GAN) with various image generation percentages. It was found that models based on Xception and VGG16 were more robust. In contrast, models based on ResNet50 and InceptionV3 were sensitive to the added machine-generated images as the accuracy degrades significantly compared to training with no artificial data.

Establishment of Priority Update Area for Land Coverage Classification Using Orthoimages and Serial Cadastral Maps

  • Song, Junyoung;Won, Taeyeon;Jo, Su Min;Eo, Yang Dam;Park, Jin Sue
    • 대한원격탐사학회지
    • /
    • 제37권4호
    • /
    • pp.763-776
    • /
    • 2021
  • This paper introduces a method of selecting priority update areas for subdivided land cover maps by training orthoimages and serial cadastral maps in a deep learning model. For the experiment, orthoimages and serial cadastral maps were obtained from the National Spatial Data Infrastructure Portal. Based on the VGG-16 model, 51,470 images were trained on 33 subdivided classifications within the experimental area and an accuracy evaluation was conducted. The overall accuracy was 61.42%. In addition, using the differences in the classification prediction probability of the misclassified polygon and the cosine similarity that numerically expresses the similarity of the land category features with the original subdivided land cover class, the cases were classified and the areas in which the boundary setting was incorrect and in which the image itself was determined to have a problem were identified as the priority update polygons that should be checked by operators.

SVM on Top of Deep Networks for Covid-19 Detection from Chest X-ray Images

  • Do, Thanh-Nghi;Le, Van-Thanh;Doan, Thi-Huong
    • Journal of information and communication convergence engineering
    • /
    • 제20권3호
    • /
    • pp.219-225
    • /
    • 2022
  • In this study, we propose training a support vector machine (SVM) model on top of deep networks for detecting Covid-19 from chest X-ray images. We started by gathering a real chest X-ray image dataset, including positive Covid-19, normal cases, and other lung diseases not caused by Covid-19. Instead of training deep networks from scratch, we fine-tuned recent pre-trained deep network models, such as DenseNet121, MobileNet v2, Inception v3, Xception, ResNet50, VGG16, and VGG19, to classify chest X-ray images into one of three classes (Covid-19, normal, and other lung). We propose training an SVM model on top of deep networks to perform a nonlinear combination of deep network outputs, improving classification over any single deep network. The empirical test results on the real chest X-ray image dataset show that deep network models, with an exception of ResNet50 with 82.44%, provide an accuracy of at least 92% on the test set. The proposed SVM on top of the deep network achieved the highest accuracy of 96.16%.

딥러닝 기반 후두부 질환 내시경 영상판독 보조기술 개발 (Development of Deep Learning-based Clinical Decision Supporting Technique for Laryngeal Disease using Endoscopic Images)

  • 정인호;황영준;성의숙;남경원
    • 대한의용생체공학회:의공학회지
    • /
    • 제43권2호
    • /
    • pp.102-108
    • /
    • 2022
  • Purpose: To propose a deep learning-based clinical decision support technique for laryngeal disease on epiglottis, tongue and vocal cords. Materials and Methods: A total of 873 laryngeal endoscopic images were acquired from the PACS database of Pusan N ational University Yangsan Hospital. and VGG16 model was applied with transfer learning and fine-tuning. Results: The values of precision, recall, accuracy and F1-score for test dataset were 0.94, 0.97, 0.95 and 0.95 for epiglottis images, 0.91, 1.00, 0.95 and 0.95 for tongue images, and 0.90, 0.64, 0.73 and 0.75 for vocal cord images, respectively. Conclusion: Experimental results demonstrated that the proposed model have a potential as a tool for decision-supporting of otolaryngologist during manual inspection of laryngeal endoscopic images.

A Defect Detection Algorithm of Denim Fabric Based on Cascading Feature Extraction Architecture

  • Shuangbao, Ma;Renchao, Zhang;Yujie, Dong;Yuhui, Feng;Guoqin, Zhang
    • Journal of Information Processing Systems
    • /
    • 제19권1호
    • /
    • pp.109-117
    • /
    • 2023
  • Defect detection is one of the key factors in fabric quality control. To improve the speed and accuracy of denim fabric defect detection, this paper proposes a defect detection algorithm based on cascading feature extraction architecture. Firstly, this paper extracts these weight parameters of the pre-trained VGG16 model on the large dataset ImageNet and uses its portability to train the defect detection classifier and the defect recognition classifier respectively. Secondly, retraining and adjusting partial weight parameters of the convolution layer were retrained and adjusted from of these two training models on the high-definition fabric defect dataset. The last step is merging these two models to get the defect detection algorithm based on cascading architecture. Then there are two comparative experiments between this improved defect detection algorithm and other feature extraction methods, such as VGG16, ResNet-50, and Xception. The results of experiments show that the defect detection accuracy of this defect detection algorithm can reach 94.3% and the speed is also increased by 1-3 percentage points.

Malware Classification using Dynamic Analysis with Deep Learning

  • Asad Amin;Muhammad Nauman Durrani;Nadeem Kafi;Fahad Samad;Abdul Aziz
    • International Journal of Computer Science & Network Security
    • /
    • 제23권8호
    • /
    • pp.49-62
    • /
    • 2023
  • There has been a rapid increase in the creation and alteration of new malware samples which is a huge financial risk for many organizations. There is a huge demand for improvement in classification and detection mechanisms available today, as some of the old strategies like classification using mac learning algorithms were proved to be useful but cannot perform well in the scalable auto feature extraction scenario. To overcome this there must be a mechanism to automatically analyze malware based on the automatic feature extraction process. For this purpose, the dynamic analysis of real malware executable files has been done to extract useful features like API call sequence and opcode sequence. The use of different hashing techniques has been analyzed to further generate images and convert them into image representable form which will allow us to use more advanced classification approaches to classify huge amounts of images using deep learning approaches. The use of deep learning algorithms like convolutional neural networks enables the classification of malware by converting it into images. These images when fed into the CNN after being converted into the grayscale image will perform comparatively well in case of dynamic changes in malware code as image samples will be changed by few pixels when classified based on a greyscale image. In this work, we used VGG-16 architecture of CNN for experimentation.

악성코드 이미지화와 전이학습을 이용한 악성코드 분류 기법 (Malware Classification Method using Malware Visualization and Transfer Learning)

  • 이종관;이민우
    • 한국정보통신학회:학술대회논문집
    • /
    • 한국정보통신학회 2021년도 춘계학술대회
    • /
    • pp.555-556
    • /
    • 2021
  • 본 논문은 악성코드의 이미지화와 전이학습을 이용한 악성코드 분류 방안을 제안한다. 공개된 악성코드는 쉽게 재사용 또는 변형이 가능하다. 그런데 전통적인 악성코드 탐지 기법은 변형된 악성코드를 탐지하는데 취약하다. 동일한 부류에 속하는 악성코드들은 서로 유사한 이미지로 변환된다. 따라서 제안하는 기법은 악성코드를 이미지화하고 이미지 분류 분야에서 검증된 딥러닝 모델을 사용하여 악성코드의 부류를 분류한다. Malimg 데이터셋에 대해 VGG-16 모델을 이용하여 실험한 결과 98% 이상의 분류 정확도를 나타냈다.

  • PDF

비정형 데이터와 딥러닝을 활용한 내수침수 탐지기술 개발 (Development of a method for urban flooding detection using unstructured data and deep learing)

  • 이하늘;김형수;김수전;김동현;김종성
    • 한국수자원학회논문집
    • /
    • 제54권12호
    • /
    • pp.1233-1242
    • /
    • 2021
  • 본 연구에서는 비정형 데이터인 사진자료를 이용하여 침수의 발생여부를 판단하는 모델을 개발하였다. 침수분류를 모델 개발을 위하여 CNN기반의 VGG16, VGG19을 이용하였다. 모델을 개발하기 위하여 침수사진과 침수가 발생하지 않은 사진을 웹크롤링 방법을 이용하여 사진을 수집하였다. 웹크롤링 방법을 이용하여 수집한 데이터는 노이즈 데이터가 포함되어 있기 때문에 1차적으로 본 연구와 상관없는 데이터는 소거하였으며, 2차적으로 모델 적용을 위하여 224 × 224로 사진 사이즈를 일괄 변경하였다. 또한 사진의 다양성을 위해서 사진의 각도를 변환하여 이미지 증식을 수행하였으며. 최종적으로 침수사진 2,500장과 침수가 발생하지 않은 사진 2,500장을 이용하여 학습을 수행하였다. 모델 평가결과 모델의 평균 분류성능은 97%로 나타났으며. 향후 본 연구결과를 통하여 개발된 모델을 CCTV관제센터 시스템에 탑재한다면 신속하게 침수피해에 대한 대처가 이루어 질 수 있을 것이라 판단된다.

공연예술에서 광고포스터의 이미지 특성을 활용한 딥러닝 기반 관객예측 (Deep Learning-Based Box Office Prediction Using the Image Characteristics of Advertising Posters in Performing Arts)

  • 조유정;강경표;권오병
    • 한국전자거래학회지
    • /
    • 제26권2호
    • /
    • pp.19-43
    • /
    • 2021
  • 공연예술 기관에서의 공연에 대한 흥행 예측은 공연예술 산업 및 기관에서 매우 흥미롭고도 중요한 문제이다. 이를 위해 출연진, 공연장소, 가격 등 정형화된 데이터를 활용한 전통적인 예측방법론, 데이터마이닝 방법론이 제시되어 왔다. 그런데 관객들은 공연안내 포스터에 의하여 관람 의도가 소구되는 경향이 있음에도 불구하고, 포스터 이미지 분석을 통한 흥행 예측은 거의 시도되지 않았다. 그러나 최근 이미지를 통해 판별하는 CNN 계열의 딥러닝 방법이 개발되면서 포스터 분석의 가능성이 열렸다. 이에 본 연구의 목적은 공연 관련 포스터 이미지를 통해 흥행을 예측할 수 있는 딥러닝 방법을 제안하는 것이다. 이를 위해 KOPIS 공연예술 통합전산망에 공개된 포스터 이미지를 학습데이터로 하여 Pure CNN, VGG-16, Inception-v3, ResNet50 등 딥러닝 알고리즘을 통해 예측을 수행하였다. 또한 공연 관련 정형데이터를 활용한 전통적 회귀분석 방법론과의 앙상블을 시도하였다. 그 결과 흥행 예측 정확도 85%를 상회하는 높은 판별 성과를 보였다. 본 연구는 공연예술 분야에서 이미지 정보를 활용하여 흥행을 예측하는 첫 시도이며 본 연구에서 제안한 방법은 연극 외에 영화, 기관 홍보, 기업 제품 광고 등 포스터 기반의 광고를 하는 영역으로도 적용이 가능할 것이다.