• Title/Summary/Keyword: Image Deep Learning

Search Result 1,827, Processing Time 0.027 seconds

Bio-Cell Image Segmentation based on Deep Learning using Denoising Autoencoder and Graph Cuts (디노이징 오토인코더와 그래프 컷을 이용한 딥러닝 기반 바이오-셀 영상 분할)

  • Lim, Seon-Ja;Vununu, Caleb;Kwon, Oh-Heum;Lee, Suk-Hwan;Kwon, Ki-Ryoug
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.10
    • /
    • pp.1326-1335
    • /
    • 2021
  • As part of the cell division method, we proposed a method for segmenting images generated by topography microscopes through deep learning-based feature generation and graph segmentation. Hybrid vector shapes preserve the overall shape and boundary information of cells, so most cell shapes can be captured without any post-processing burden. NIH-3T3 and Hela-S3 cells have satisfactory results in cell description preservation. Compared to other deep learning methods, the proposed cell image segmentation method does not require postprocessing. It is also effective in preserving the overall morphology of cells and has shown better results in terms of cell boundary preservation.

A Deep Learning Model for Predicting User Personality Using Social Media Profile Images

  • Kanchana, T.S.;Zoraida, B.S.E.
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.11
    • /
    • pp.265-271
    • /
    • 2022
  • Social media is a form of communication based on the internet to share information through content and images. Their choice of profile images and type of image they post can be closely connected to their personality. The user posted images are designated as personality traits. The objective of this study is to predict five factor model personality dimensions from profile images by using deep learning and neural networks. Developed a deep learning framework-based neural network for personality prediction. The personality types of the Big Five Factor model can be quantified from user profile images. To measure the effectiveness, proposed two models using convolution Neural Networks to classify each personality of the user. Done performance analysis among two different models for efficiently predict personality traits from profile image. It was found that VGG-69 CNN models are best performing models for producing the classification accuracy of 91% to predict user personality traits.

Development of a Deep Learning Algorithm for Small Object Detection in Real-Time (실시간 기반 매우 작은 객체 탐지를 위한 딥러닝 알고리즘 개발)

  • Wooseong Yeo;Meeyoung Park
    • Journal of the Korean Society of Industry Convergence
    • /
    • v.27 no.4_2
    • /
    • pp.1001-1007
    • /
    • 2024
  • Recent deep learning algorithms for object detection in real-time play a crucial role in various applications such as autonomous driving, traffic monitoring, health care, and water quality monitoring. The size of small objects, in particular, significantly impacts the accuracy of detection models. However, data containing small objects can lead to underfitting issues in models. Therefore, this study developed a deep learning model capable of quickly detecting small objects to provide more accurate predictions. The RE-SOD (Residual block based Small Object Detector) developed in this research enhances the detection performance for small objects by using RGB separation preprocessing and residual blocks. The model achieved an accuracy of 1.0 in image classification and an mAP50-95 score of 0.944 in object detection. The performance of this model was validated by comparing it with real-time detection models such as YOLOv5, YOLOv7, and YOLOv8.

A Study on Flame Detection using Faster R-CNN and Image Augmentation Techniques (Faster R-CNN과 이미지 오그멘테이션 기법을 이용한 화염감지에 관한 연구)

  • Kim, Jae-Jung;Ryu, Jin-Kyu;Kwak, Dong-Kurl;Byun, Sun-Joon
    • Journal of IKEEE
    • /
    • v.22 no.4
    • /
    • pp.1079-1087
    • /
    • 2018
  • Recently, computer vision field based deep learning artificial intelligence has become a hot topic among various image analysis boundaries. In this study, flames are detected in fire images using the Faster R-CNN algorithm, which is used to detect objects within the image, among various image recognition algorithms based on deep learning. In order to improve fire detection accuracy through a small amount of data sets in the learning process, we use image augmentation techniques, and learn image augmentation by dividing into 6 types and compare accuracy, precision and detection rate. As a result, the detection rate increases as the type of image augmentation increases. However, as with the general accuracy and detection rate of other object detection models, the false detection rate is also increased from 10% to 30%.

Deep Convolution Neural Networks in Computer Vision: a Review

  • Yoo, Hyeon-Joong
    • IEIE Transactions on Smart Processing and Computing
    • /
    • v.4 no.1
    • /
    • pp.35-43
    • /
    • 2015
  • Over the past couple of years, tremendous progress has been made in applying deep learning (DL) techniques to computer vision. Especially, deep convolutional neural networks (DCNNs) have achieved state-of-the-art performance on standard recognition datasets and tasks such as ImageNet Large-Scale Visual Recognition Challenge (ILSVRC). Among them, GoogLeNet network which is a radically redesigned DCNN based on the Hebbian principle and scale invariance set the new state of the art for classification and detection in the ILSVRC 2014. Since there exist various deep learning techniques, this review paper is focusing on techniques directly related to DCNNs, especially those needed to understand the architecture and techniques employed in GoogLeNet network.

A Study on Image Labeling Technique for Deep-Learning-Based Multinational Tanks Detection Model

  • Kim, Taehoon;Lim, Dongkyun
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.14 no.4
    • /
    • pp.58-63
    • /
    • 2022
  • Recently, the improvement of computational processing ability due to the rapid development of computing technology has greatly advanced the field of artificial intelligence, and research to apply it in various domains is active. In particular, in the national defense field, attention is paid to intelligent recognition among machine learning techniques, and efforts are being made to develop object identification and monitoring systems using artificial intelligence. To this end, various image processing technologies and object identification algorithms are applied to create a model that can identify friendly and enemy weapon systems and personnel in real-time. In this paper, we conducted image processing and object identification focused on tanks among various weapon systems. We initially conducted processing the tanks' image using a convolutional neural network, a deep learning technique. The feature map was examined and the important characteristics of the tanks crucial for learning were derived. Then, using YOLOv5 Network, a CNN-based object detection network, a model trained by labeling the entire tank and a model trained by labeling only the turret of the tank were created and the results were compared. The model and labeling technique we proposed in this paper can more accurately identify the type of tank and contribute to the intelligent recognition system to be developed in the future.

Ensemble Deep Network for Dense Vehicle Detection in Large Image

  • Yu, Jae-Hyoung;Han, Youngjoon;Kim, JongKuk;Hahn, Hernsoo
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.1
    • /
    • pp.45-55
    • /
    • 2021
  • This paper has proposed an algorithm that detecting for dense small vehicle in large image efficiently. It is consisted of two Ensemble Deep-Learning Network algorithms based on Coarse to Fine method. The system can detect vehicle exactly on selected sub image. In the Coarse step, it can make Voting Space using the result of various Deep-Learning Network individually. To select sub-region, it makes Voting Map by to combine each Voting Space. In the Fine step, the sub-region selected in the Coarse step is transferred to final Deep-Learning Network. The sub-region can be defined by using dynamic windows. In this paper, pre-defined mapping table has used to define dynamic windows for perspective road image. Identity judgment of vehicle moving on each sub-region is determined by closest center point of bottom of the detected vehicle's box information. And it is tracked by vehicle's box information on the continuous images. The proposed algorithm has evaluated for performance of detection and cost in real time using day and night images captured by CCTV on the road.

Evaluation of Deep Learning Model for Scoliosis Pre-Screening Using Preprocessed Chest X-ray Images

  • Min Gu Jang;Jin Woong Yi;Hyun Ju Lee;Ki Sik Tae
    • Journal of Biomedical Engineering Research
    • /
    • v.44 no.4
    • /
    • pp.293-301
    • /
    • 2023
  • Scoliosis is a three-dimensional deformation of the spine that is a deformity induced by physical or disease-related causes as the spine is rotated abnormally. Early detection has a significant influence on the possibility of nonsurgical treatment. To train a deep learning model with preprocessed images and to evaluate the results with and without data augmentation to enable the diagnosis of scoliosis based only on a chest X-ray image. The preprocessed images in which only the spine, rib contours, and some hard tissues were left from the original chest image, were used for learning along with the original images, and three CNN(Convolutional Neural Networks) models (VGG16, ResNet152, and EfficientNet) were selected to proceed with training. The results obtained by training with the preprocessed images showed a superior accuracy to those obtained by training with the original image. When the scoliosis image was added through data augmentation, the accuracy was further improved, ultimately achieving a classification accuracy of 93.56% with the ResNet152 model using test data. Through supplementation with future research, the method proposed herein is expected to allow the early diagnosis of scoliosis as well as cost reduction by reducing the burden of additional radiographic imaging for disease detection.

Ensemble UNet 3+ for Medical Image Segmentation

  • JongJin, Park
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.1
    • /
    • pp.269-274
    • /
    • 2023
  • In this paper, we proposed a new UNet 3+ model for medical image segmentation. The proposed ensemble(E) UNet 3+ model consists of UNet 3+s of varying depths into one unified architecture. UNet 3+s of varying depths have same encoder, but have their own decoders. They can bridge semantic gap between encoder and decoder nodes of UNet 3+. Deep supervision was used for learning on a total of 8 nodes of the E-UNet 3+ to improve performance. The proposed E-UNet 3+ model shows better segmentation results than those of the UNet 3+. As a result of the simulation, the E-UNet 3+ model using deep supervision was the best with loss function values of 0.8904 and 0.8562 for training and validation data. For the test data, the UNet 3+ model using deep supervision was the best with a value of 0.7406. Qualitative comparison of the simulation results shows the results of the proposed model are better than those of existing UNet 3+.

Facial Expression Recognition through Self-supervised Learning for Predicting Face Image Sequence

  • Yoon, Yeo-Chan;Kim, Soo Kyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.9
    • /
    • pp.41-47
    • /
    • 2022
  • In this paper, we propose a new and simple self-supervised learning method that predicts the middle image of a face image sequence for automatic expression recognition. Automatic facial expression recognition can achieve high performance through deep learning methods, however, generally requires a expensive large data set. The size of the data set and the performance of the algorithm are tend to be proportional. The proposed method learns latent deep representation of a face through self-supervised learning using an existing dataset without constructing an additional dataset. Then it transfers the learned parameter to new facial expression reorganization model for improving the performance of automatic expression recognition. The proposed method showed high performance improvement for two datasets, CK+ and AFEW 8.0, and showed that the proposed method can achieve a great effect.