• Title/Summary/Keyword: Deep Learning Dataset

Search Result 815, Processing Time 0.026 seconds

Abnormal Electrocardiogram Signal Detection Based on the BiLSTM Network

  • Asif, Husnain;Choe, Tae-Young
    • International Journal of Contents
    • /
    • v.18 no.2
    • /
    • pp.68-80
    • /
    • 2022
  • The health of the human heart is commonly measured using ECG (Electrocardiography) signals. To identify any anomaly in the human heart, the time-sequence of ECG signals is examined manually by a cardiologist or cardiac electrophysiologist. Lightweight anomaly detection on ECG signals in an embedded system is expected to be popular in the near future, because of the increasing number of heart disease symptoms. Some previous research uses deep learning networks such as LSTM and BiLSTM to detect anomaly signals without any handcrafted feature. Unfortunately, lightweight LSTMs show low precision and heavy LSTMs require heavy computing powers and volumes of labeled dataset for symptom classification. This paper proposes an ECG anomaly detection system based on two level BiLSTM for acceptable precision with lightweight networks, which is lightweight and usable at home. Also, this paper presents a new threshold technique which considers statistics of the current ECG pattern. This paper's proposed model with BiLSTM detects ECG signal anomaly in 0.467 ~ 1.0 F1 score, compared to 0.426 ~ 0.978 F1 score of the similar model with LSTM except one highly noisy dataset.

High-Resolution Satellite Image Super-Resolution Using Image Degradation Model with MTF-Based Filters

  • Minkyung Chung;Minyoung Jung;Yongil Kim
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.4
    • /
    • pp.395-407
    • /
    • 2023
  • Super-resolution (SR) has great significance in image processing because it enables downstream vision tasks with high spatial resolution. Recently, SR studies have adopted deep learning networks and achieved remarkable SR performance compared to conventional example-based methods. Deep-learning-based SR models generally require low-resolution (LR) images and the corresponding high-resolution (HR) images as training dataset. Due to the difficulties in obtaining real-world LR-HR datasets, most SR models have used only HR images and generated LR images with predefined degradation such as bicubic downsampling. However, SR models trained on simple image degradation do not reflect the properties of the images and often result in deteriorated SR qualities when applied to real-world images. In this study, we propose an image degradation model for HR satellite images based on the modulation transfer function (MTF) of an imaging sensor. Because the proposed method determines the image degradation based on the sensor properties, it is more suitable for training SR models on remote sensing images. Experimental results on HR satellite image datasets demonstrated the effectiveness of applying MTF-based filters to construct a more realistic LR-HR training dataset.

A Framework for Facial Expression Recognition Combining Contextual Information and Attention Mechanism

  • Jianzeng Chen;Ningning Chen
    • Journal of Information Processing Systems
    • /
    • v.20 no.4
    • /
    • pp.535-549
    • /
    • 2024
  • Facial expressions (FEs) serve as fundamental components for human emotion assessment and human-computer interaction. Traditional convolutional neural networks tend to overlook valuable information during the FE feature extraction, resulting in suboptimal recognition rates. To address this problem, we propose a deep learning framework that incorporates hierarchical feature fusion, contextual data, and an attention mechanism for precise FE recognition. In our approach, we leveraged an enhanced VGGNet16 as the backbone network and introduced an improved group convolutional channel attention (GCCA) module in each block to emphasize the crucial expression features. A partial decoder was added at the end of the backbone network to facilitate the fusion of multilevel features for a comprehensive feature map. A reverse attention mechanism guides the model to refine details layer-by-layer while introducing contextual information and extracting richer expression features. To enhance feature distinguishability, we employed islanding loss in combination with softmax loss, creating a joint loss function. Using two open datasets, our experimental results demonstrated the effectiveness of our framework. Our framework achieved an average accuracy rate of 74.08% on the FER2013 dataset and 98.66% on the CK+ dataset, outperforming advanced methods in both recognition accuracy and stability.

Development of Security Anomaly Detection Algorithms using Machine Learning (기계 학습을 활용한 보안 이상징후 식별 알고리즘 개발)

  • Hwangbo, Hyunwoo;Kim, Jae Kyung
    • The Journal of Society for e-Business Studies
    • /
    • v.27 no.1
    • /
    • pp.1-13
    • /
    • 2022
  • With the development of network technologies, the security to protect organizational resources from internal and external intrusions and threats becomes more important. Therefore in recent years, the anomaly detection algorithm that detects and prevents security threats with respect to various security log events has been actively studied. Security anomaly detection algorithms that have been developed based on rule-based or statistical learning in the past are gradually evolving into modeling based on machine learning and deep learning. In this study, we propose a deep-autoencoder model that transforms LSTM-autoencoder as an optimal algorithm to detect insider threats in advance using various machine learning analysis methodologies. This study has academic significance in that it improved the possibility of adaptive security through the development of an anomaly detection algorithm based on unsupervised learning, and reduced the false positive rate compared to the existing algorithm through supervised true positive labeling.

A Study of Kernel Characteristics of CNN Deep Learning for Effective Fire Detection Based on Video (영상기반의 화재 검출에 효과적인 CNN 심층학습의 커널 특성에 대한 연구)

  • Son, Geum-Young;Park, Jang-Sik
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.6
    • /
    • pp.1257-1262
    • /
    • 2018
  • In this paper, a deep learning method is proposed to detect the fire effectively by using video of surveillance camera. Based on AlexNet model, classification performance is compared according to kernel size and stride of convolution layer. Dataset for learning and interfering are classified into two classes such as normal and fire. Normal images include clouds, and foggy images, and fire images include smoke and flames images, respectively. As results of simulations, it is shown that the larger kernel size and smaller stride shows better performance.

Development of Deep Learning-based Clinical Decision Supporting Technique for Laryngeal Disease using Endoscopic Images (딥러닝 기반 후두부 질환 내시경 영상판독 보조기술 개발)

  • Jung, In Ho;Hwang, Young Jun;Sung, Eui-Suk;Nam, Kyoung Won
    • Journal of Biomedical Engineering Research
    • /
    • v.43 no.2
    • /
    • pp.102-108
    • /
    • 2022
  • Purpose: To propose a deep learning-based clinical decision support technique for laryngeal disease on epiglottis, tongue and vocal cords. Materials and Methods: A total of 873 laryngeal endoscopic images were acquired from the PACS database of Pusan N ational University Yangsan Hospital. and VGG16 model was applied with transfer learning and fine-tuning. Results: The values of precision, recall, accuracy and F1-score for test dataset were 0.94, 0.97, 0.95 and 0.95 for epiglottis images, 0.91, 1.00, 0.95 and 0.95 for tongue images, and 0.90, 0.64, 0.73 and 0.75 for vocal cord images, respectively. Conclusion: Experimental results demonstrated that the proposed model have a potential as a tool for decision-supporting of otolaryngologist during manual inspection of laryngeal endoscopic images.

Optimized Deep Learning Techniques for Disease Detection in Rice Crop using Merged Datasets

  • Muhammad Junaid;Sohail Jabbar;Muhammad Munwar Iqbal;Saqib Majeed;Mubarak Albathan;Qaisar Abbas;Ayyaz Hussain
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.3
    • /
    • pp.57-66
    • /
    • 2023
  • Rice is an important food crop for most of the population in the world and it is largely cultivated in Pakistan. It not only fulfills food demand in the country but also contributes to the wealth of Pakistan. But its production can be affected by climate change. The irregularities in the climate can cause several diseases such as brown spots, bacterial blight, tungro and leaf blasts, etc. Detection of these diseases is necessary for suitable treatment. These diseases can be effectively detected using deep learning such as Convolution Neural networks. Due to the small dataset, transfer learning models such as vgg16 model can effectively detect the diseases. In this paper, vgg16, inception and xception models are used. Vgg16, inception and xception models have achieved 99.22%, 88.48% and 93.92% validation accuracies when the epoch value is set to 10. Evaluation of models has also been done using accuracy, recall, precision, and confusion matrix.

Detection of Bacteria in Blood in Darkfield Microscopy Image (암시야 현미경 영상에서 혈액 내 박테리아 검출 방법)

  • Park, Hyun-jun
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.183-185
    • /
    • 2021
  • Detecting bacteria in blood could be an important research area in medicine and computer vision. In this paper, we propose a method for detecting bacteria in blood from 366 darkfield microscopy images acquired at Kaggle. Generate a training dataset through preprocessing and data augmentation using image processing techniques, and define a deep learning model for learning it. As a result of the experiment, it was confirmed that the proposed deep learning model effectively detects red blood cells and bacteria in darkfield microscopy images. In this paper, we learned using a relatively simple model, but it seems that more accurate results can be obtained by using a deeper model.

  • PDF

Auto Labelling System using Object Segmentation Technology (객체 분할 기법을 활용한 자동 라벨링 구축)

  • Moon, Jun-hwi;Park, Seong-hyeon;Choi, Jiyoung;Shin, Wonsun;Jung, Heokyung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.222-224
    • /
    • 2022
  • Deep learning-based computer vision applications in the field of object segmentation take a transfer learning method using hyperparameters and models pretrained and distributed by STOA techniques to improve performance. Custom datasets used in this process require a lot of resources, such as time and labeling, in labeling tasks to generate Ground Truth information. In this paper, we present an automatic labeling construction method using object segmentation techniques so that resources such as time and labeling can be used less to build custom datasets used in deep learning neural networks.

  • PDF

Animal Face Classification using Dual Deep Convolutional Neural Network

  • Khan, Rafiul Hasan;Kang, Kyung-Won;Lim, Seon-Ja;Youn, Sung-Dae;Kwon, Oh-Jun;Lee, Suk-Hwan;Kwon, Ki-Ryong
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.4
    • /
    • pp.525-538
    • /
    • 2020
  • A practical animal face classification system that classifies animals in image and video data is considered as a pivotal topic in machine learning. In this research, we are proposing a novel method of fully connected dual Deep Convolutional Neural Network (DCNN), which extracts and analyzes image features on a large scale. With the inclusion of the state of the art Batch Normalization layer and Exponential Linear Unit (ELU) layer, our proposed DCNN has gained the capability of analyzing a large amount of dataset as well as extracting more features than before. For this research, we have built our dataset containing ten thousand animal faces of ten animal classes and a dual DCNN. The significance of our network is that it has four sets of convolutional functions that work laterally with each other. We used a relatively small amount of batch size and a large number of iteration to mitigate overfitting during the training session. We have also used image augmentation to vary the shapes of the training images for the better learning process. The results demonstrate that, with an accuracy rate of 92.0%, the proposed DCNN outruns its counterparts while causing less computing costs.