• 제목/요약/키워드: Fully Convolutional Layer

검색결과 56건 처리시간 0.02초

폐 결절 검출을 위한 합성곱 신경망의 성능 개선 (Performance Improvement of Convolutional Neural Network for Pulmonary Nodule Detection)

  • 김한웅;김병남;이지은;장원석;유선국
    • 대한의용생체공학회:의공학회지
    • /
    • 제38권5호
    • /
    • pp.237-241
    • /
    • 2017
  • Early detection of the pulmonary nodule is important for diagnosis and treatment of lung cancer. Recently, CT has been used as a screening tool for lung nodule detection. And, it has been reported that computer aided detection(CAD) systems can improve the accuracy of the radiologist in detection nodules on CT scan. The previous study has been proposed a method using Convolutional Neural Network(CNN) in Lung CAD system. But the proposed model has a limitation in accuracy due to its sparse layer structure. Therefore, we propose a Deep Convolutional Neural Network to overcome this limitation. The model proposed in this work is consist of 14 layers including 8 convolutional layers and 4 fully connected layers. The CNN model is trained and tested with 61,404 regions-of-interest (ROIs) patches of lung image including 39,760 nodules and 21,644 non-nodules extracted from the Lung Image Database Consortium(LIDC) dataset. We could obtain the classification accuracy of 91.79% with the CNN model presented in this work. To prevent overfitting, we trained the model with Augmented Dataset and regularization term in the cost function. With L1, L2 regularization at Training process, we obtained 92.39%, 92.52% of accuracy respectively. And we obtained 93.52% with data augmentation. In conclusion, we could obtain the accuracy of 93.75% with L2 Regularization and Data Augmentation.

딥러닝 기반 3차원 라이다의 반사율 세기 신호를 이용한 흑백 영상 생성 기법 (Deep Learning Based Gray Image Generation from 3D LiDAR Reflection Intensity)

  • 김현구;유국열;박주현;정호열
    • 대한임베디드공학회논문지
    • /
    • 제14권1호
    • /
    • pp.1-9
    • /
    • 2019
  • In this paper, we propose a method of generating a 2D gray image from LiDAR 3D reflection intensity. The proposed method uses the Fully Convolutional Network (FCN) to generate the gray image from 2D reflection intensity which is projected from LiDAR 3D intensity. Both encoder and decoder of FCN are configured with several convolution blocks in the symmetric fashion. Each convolution block consists of a convolution layer with $3{\times}3$ filter, batch normalization layer and activation function. The performance of the proposed method architecture is empirically evaluated by varying depths of convolution blocks. The well-known KITTI data set for various scenarios is used for training and performance evaluation. The simulation results show that the proposed method produces the improvements of 8.56 dB in peak signal-to-noise ratio and 0.33 in structural similarity index measure compared with conventional interpolation methods such as inverse distance weighted and nearest neighbor. The proposed method can be possibly used as an assistance tool in the night-time driving system for autonomous vehicles.

A Deep Approach for Classifying Artistic Media from Artworks

  • Yang, Heekyung;Min, Kyungha
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제13권5호
    • /
    • pp.2558-2573
    • /
    • 2019
  • We present a deep CNN-based approach for classifying artistic media from artwork images. We aim to classify most frequently used artistic media including oilpaint brush, watercolor brush, pencil and pastel, etc. For this purpose, we extend VGGNet, one of the most widely used CNN structure, by substituting its last layer with a fully convolutional layer, which reveals class activation map (CAM), the region of classification. We build two artwork image datasets: YMSet that collects more than 4K artwork images for four most frequently used artistic media from various internet websites and WikiSet that collects almost 9K artwork images for ten most frequently used media from WikiArt. We execute a human baseline experiment to compare the classification performance. Through our experiments, we conclude that our classifier is superior in classifying artistic media to human.

CCTV를 이용한 터널내 사고감지 시스템 (Accident Detection System in Tunnel using CCTV)

  • 이세훈;이승엽;노영훈
    • 한국컴퓨터정보학회:학술대회논문집
    • /
    • 한국컴퓨터정보학회 2021년도 제64차 하계학술대회논문집 29권2호
    • /
    • pp.3-4
    • /
    • 2021
  • 폐쇄된 터널 내부에서는 사고가 일어날 경우 외부에서는 터널 내 상황을 알 수가 없어 경미한 사고라 하더라도 대형 후속 2차 사고로 이어질 가능성이 크다. 또한영상탐지로사고 상황의 오검출을 줄이기 위해서, 본 연구에서는기존의 많은 CNN 모델 중 보유한 데이터에 가장 적합한 모델을 선택하는 과정에서 가장 좋은 성능을 보인 VGG16 모델을 전이학습 시키고 fully connected layer의 일부 layer에 Dropout을 적용시켜 Overfitting을일부방지하는 CNN 모델을 생성한 뒤Yolo를 이용한 영상 내 객체인식, OpenCV를 이용한 영상 프레임 내에서 객체의ROI를 추출하고이를 CNN 모델과 비교하여오검출을 줄이면서 사고를 검출하는 시스템을 제안하였다.

  • PDF

CNN을 사용한 차선검출 시스템 (Lane Detection System using CNN)

  • 김지훈;이대식;이민호
    • 대한임베디드공학회논문지
    • /
    • 제11권3호
    • /
    • pp.163-171
    • /
    • 2016
  • Lane detection is a widely researched topic. Although simple road detection is easily achieved by previous methods, lane detection becomes very difficult in several complex cases involving noisy edges. To address this, we use a Convolution neural network (CNN) for image enhancement. CNN is a deep learning method that has been very successfully applied in object detection and recognition. In this paper, we introduce a robust lane detection method based on a CNN combined with random sample consensus (RANSAC) algorithm. Initially, we calculate edges in an image using a hat shaped kernel, then we detect lanes using the CNN combined with the RANSAC. In the training process of the CNN, input data consists of edge images and target data is images that have real white color lanes on an otherwise black background. The CNN structure consists of 8 layers with 3 convolutional layers, 2 subsampling layers and multi-layer perceptron (MLP) of 3 fully-connected layers. Convolutional and subsampling layers are hierarchically arranged to form a deep structure. Our proposed lane detection algorithm successfully eliminates noise lines and was found to perform better than other formal line detection algorithms such as RANSAC

Human Activity Recognition Based on 3D Residual Dense Network

  • Park, Jin-Ho;Lee, Eung-Joo
    • 한국멀티미디어학회논문지
    • /
    • 제23권12호
    • /
    • pp.1540-1551
    • /
    • 2020
  • Aiming at the problem that the existing human behavior recognition algorithm cannot fully utilize the multi-level spatio-temporal information of the network, a human behavior recognition algorithm based on a dense three-dimensional residual network is proposed. First, the proposed algorithm uses a dense block of three-dimensional residuals as the basic module of the network. The module extracts the hierarchical features of human behavior through densely connected convolutional layers; Secondly, the local feature aggregation adaptive method is used to learn the local dense features of human behavior; Then, the residual connection module is applied to promote the flow of feature information and reduced the difficulty of training; Finally, the multi-layer local feature extraction of the network is realized by cascading multiple three-dimensional residual dense blocks, and use the global feature aggregation adaptive method to learn the features of all network layers to realize human behavior recognition. A large number of experimental results on benchmark datasets KTH show that the recognition rate (top-l accuracy) of the proposed algorithm reaches 93.52%. Compared with the three-dimensional convolutional neural network (C3D) algorithm, it has improved by 3.93 percentage points. The proposed algorithm framework has good robustness and transfer learning ability, and can effectively handle a variety of video behavior recognition tasks.

전이학습과 딥러닝 네트워크를 활용한 고해상도 위성영상의 변화탐지 (Change Detection for High-resolution Satellite Images Using Transfer Learning and Deep Learning Network)

  • 송아람;최재완;김용일
    • 한국측량학회지
    • /
    • 제37권3호
    • /
    • pp.199-208
    • /
    • 2019
  • 운용 가능한 위성의 수가 증가하고 기술이 진보함에 따라 영상정보의 성과물이 다양해지고 많은 양의 자료가 축적되고 있다. 본 연구에서는 기구축된 영상정보를 활용하여 부족한 훈련자료의 문제를 극복하고 딥러닝(deep learning) 기법의 장점을 활용하고자 전이학습과 변화탐지 네트워크를 활용한 고해상도 위성영상의 변화탐지를 수행하였다. 본 연구에서 활용한 딥러닝 네트워크는 공간 및 분광 정보를 추출하는 합성곱 레이어(convolutional layer)와 시계열 정보를 분석하는 합성곱 장단기 메모리 레이어(convolutional long short term memory layer)로 구성되었으며, 고해상도 다중분광 영상에 최적화된 정보를 추출하기 위하여 커널(kernel)의 차원에 따른 정확도를 비교하였다. 또한, 학습된 커널 정보를 활용하기 위하여 변화탐지 네트워크의 초기 합성곱 레이어를 고해상도 항공영상인 ISPRS (International Society for Photogrammetry and Remote Sensing) 데이터셋에서 추출된 40,000개의 패치로 학습된 값으로 초기화하였다. 다시기 KOMPSAT-3A (KOrean Multi-Purpose SATllite-3A) 영상에 대한 실험 결과, 전이학습과 딥러닝 네트워크를 활용할 경우 기복 변위 및 그림자 등으로 인한 변화에 덜 민감하게 반응하며 분류 항목이 달라진 지역의 변화를 보다 효과적으로 추출할 수 있었으며, 2차원 커널보다 3차원 커널을 사용할 때 변화탐지의 정확도가 높았다. 3차원 커널은 공간 및 분광정보를 모두 고려하여 특징 맵(feature map)을 추출하기 때문에 고해상도 영상의 분류뿐만 아니라 변화탐지에도 효과적인 것을 확인하였다. 본 연구에서는 고해상도 위성영상의 변화탐지를 위한 전이학습과 딥러닝 기법의 활용 가능성을 제시하였으며, 추후 훈련된 변화탐지 네트워크를 새롭게 취득된 영상에 적용하는 연구를 수행하여 제안기법의 활용범위를 확장할 예정이다.

전이학습 기반 CNN을 통한 풀림 방지 코팅 볼트 이진 분류에 관한 연구 (Binary classification of bolts with anti-loosening coating using transfer learning-based CNN)

  • 노은솔;이사랑;홍석무
    • 한국산학기술학회논문지
    • /
    • 제22권2호
    • /
    • pp.651-658
    • /
    • 2021
  • 풀림 방지 코팅 볼트는 주로 자동차 안전 관련 부품을 결합하는 데 사용되므로 안전성 유지를 위해 코팅 결함을 사전에 감지해야 한다. 이를 위해 이전 연구 [CNN 및 모델 시각화 기법을 사용한 코팅 볼트 불량 판별]에서는 합성곱 신경망을 사용했다. 이때 합성곱 신경망은 데이터 수가 많을수록 이미지 패턴 및 특성 분석 정확도가 증가하지만 그에 따라 학습시간이 증가한다. 또한 확보 가능한 코팅 볼트 샘플이 한정적이다. 본 연구에서는 이전 연구에 전이학습을 추가적으로 적용해 데이터 개수가 적은 경우에도 코팅 결함에 대해 정확한 분류를 하고자 한다. 전이학습을 적용할 때 학습 데이터 수와 사전 학습 데이터 ImageNet 간의 유사성을 고려해 분류층만 학습했다. 데이터 학습에는 전역 평균 풀링, 선형 서포트 벡터 머신 및 완전 연결 계층과 같은 분류층을 적용했으며, 고려한 모델 중 완전 연결 계층 방법의 분류층이 가장 높은 95% 정확도를 가진다. 추가적으로 마지막 합성곱층과 분류층을 미세 조정하면 정확도는 97%까지 향상된다. 전이학습 및 미세 조정을 이용하면 선별 정확도를 향상시킴은 물론 이전보다 학습 소요시간을 절반으로 줄일 수 있음을 보였다.

Enhanced CT-image for Covid-19 classification using ResNet 50

  • Lobna M. Abouelmagd;Manal soubhy Ali Elbelkasy
    • International Journal of Computer Science & Network Security
    • /
    • 제24권1호
    • /
    • pp.119-126
    • /
    • 2024
  • Disease caused by the coronavirus (COVID-19) is sweeping the globe. There are numerous methods for identifying this disease using a chest imaging. Computerized Tomography (CT) chest scans are used in this study to detect COVID-19 disease using a pretrain Convolutional Neural Network (CNN) ResNet50. This model is based on image dataset taken from two hospitals and used to identify Covid-19 illnesses. The pre-train CNN (ResNet50) architecture was used for feature extraction, and then fully connected layers were used for classification, yielding 97%, 96%, 96%, 96% for accuracy, precision, recall, and F1-score, respectively. When combining the feature extraction techniques with the Back Propagation Neural Network (BPNN), it produced accuracy, precision, recall, and F1-scores of 92.5%, 83%, 92%, and 87.3%. In our suggested approach, we use a preprocessing phase to improve accuracy. The image was enhanced using the Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm, which was followed by cropping the image before feature extraction with ResNet50. Finally, a fully connected layer was added for classification, with results of 99.1%, 98.7%, 99%, 98.8% in terms of accuracy, precision, recall, and F1-score.

Voting and Ensemble Schemes Based on CNN Models for Photo-Based Gender Prediction

  • Jhang, Kyoungson
    • Journal of Information Processing Systems
    • /
    • 제16권4호
    • /
    • pp.809-819
    • /
    • 2020
  • Gender prediction accuracy increases as convolutional neural network (CNN) architecture evolves. This paper compares voting and ensemble schemes to utilize the already trained five CNN models to further improve gender prediction accuracy. The majority voting usually requires odd-numbered models while the proposed softmax-based voting can utilize any number of models to improve accuracy. The ensemble of CNN models combined with one more fully-connected layer requires further tuning or training of the models combined. With experiments, it is observed that the voting or ensemble of CNN models leads to further improvement of gender prediction accuracy and that especially softmax-based voters always show better gender prediction accuracy than majority voters. Also, compared with softmax-based voters, ensemble models show a slightly better or similar accuracy with added training of the combined CNN models. Softmax-based voting can be a fast and efficient way to get better accuracy without further training since the selection of the top accuracy models among available CNN pre-trained models usually leads to similar accuracy to that of the corresponding ensemble models.