• Title/Summary/Keyword: Self- Supervised Learning

Search Result 100, Processing Time 0.028 seconds

Attentive Transfer Learning via Self-supervised Learning for Cervical Dysplasia Diagnosis

  • Chae, Jinyeong;Zimmermann, Roger;Kim, Dongho;Kim, Jihie
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.453-461
    • /
    • 2021
  • Many deep learning approaches have been studied for image classification in computer vision. However, there are not enough data to generate accurate models in medical fields, and many datasets are not annotated. This study presents a new method that can use both unlabeled and labeled data. The proposed method is applied to classify cervix images into normal versus cancerous, and we demonstrate the results. First, we use a patch self-supervised learning for training the global context of the image using an unlabeled image dataset. Second, we generate a classifier model by using the transferred knowledge from self-supervised learning. We also apply attention learning to capture the local features of the image. The combined method provides better performance than state-of-the-art approaches in accuracy and sensitivity.

Software Fault Prediction using Semi-supervised Learning Methods (세미감독형 학습 기법을 사용한 소프트웨어 결함 예측)

  • Hong, Euyseok
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.3
    • /
    • pp.127-133
    • /
    • 2019
  • Most studies of software fault prediction have been about supervised learning models that use only labeled training data. Although supervised learning usually shows high prediction performance, most development groups do not have sufficient labeled data. Unsupervised learning models that use only unlabeled data for training are difficult to build and show poor performance. Semi-supervised learning models that use both labeled data and unlabeled data can solve these problems. Self-training technique requires the fewest assumptions and constraints among semi-supervised techniques. In this paper, we implemented several models using self-training algorithms and evaluated them using Accuracy and AUC. As a result, YATSI showed the best performance.

Pretext Task Analysis for Self-Supervised Learning Application of Medical Data (의료 데이터의 자기지도학습 적용을 위한 pretext task 분석)

  • Kong, Heesan;Park, Jaehun;Kim, Kwangsu
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.38-40
    • /
    • 2021
  • Medical domain has a massive number of data records without the response value. Self-supervised learning is a suitable method for medical data since it learns pretext-task and supervision, which the model can understand the semantic representation of data without response values. However, since self-supervised learning performance depends on the expression learned by the pretext-task, it is necessary to define an appropriate Pretext-task with data feature consideration. In this paper, to actively exploit the unlabeled medical data into artificial intelligence research, experimentally find pretext-tasks that suitable for the medical data and analyze the result. We use the x-ray image dataset which is effectively utilizable for the medical domain.

  • PDF

Multimodal Supervised Contrastive Learning for Crop Disease Diagnosis (멀티 모달 지도 대조 학습을 이용한 농작물 병해 진단 예측 방법)

  • Hyunseok Lee;Doyeob Yeo;Gyu-Sung Ham;Kanghan Oh
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.285-292
    • /
    • 2023
  • With the wide spread of smart farms and the advancements in IoT technology, it is easy to obtain additional data in addition to crop images. Consequently, deep learning-based crop disease diagnosis research utilizing multimodal data has become important. This study proposes a crop disease diagnosis method using multimodal supervised contrastive learning by expanding upon the multimodal self-supervised learning. RandAugment method was used to augment crop image and time series of environment data. These augmented data passed through encoder and projection head for each modality, yielding low-dimensional features. Subsequently, the proposed multimodal supervised contrastive loss helped features from the same class get closer while pushing apart those from different classes. Following this, the pretrained model was fine-tuned for crop disease diagnosis. The visualization of t-SNE result and comparative assessments of crop disease diagnosis performance substantiate that the proposed method has superior performance than multimodal self-supervised learning.

Multivariate Masked Autoencoders for Obstructive Sleep Apnea Diagnosis

  • Thanh-Cong Do;Hyung-Jeong Yang;Hyeong-Chae Yang
    • Annual Conference of KIPS
    • /
    • 2024.10a
    • /
    • pp.539-542
    • /
    • 2024
  • Obstructive sleep apnea (OSA) is a common sleep disorder, causing disrupted sleep and reduced oxygen levels in the blood. Full-night polysomnography (PSG) and home sleep apnea tests (HSAT) may offer acceptable diagnosis results but have shown some limitations. In recent years, deep learning and data analysis methods are progressively employed on electronic health records, and various methods have been developed for OSA event detection. Self-supervised learning has shown some advantages over supervised training methods, by learning more generalized feature representation of data. Unlike text or image processing, the high information density in multivariate time-series data makes it more challenging to utilize self-supervised learning. In this research, based on the characteristics of sleep dataset, we propose a self-supervised approach with Masked Autoencoder, which masks portions of the input and attempting to reconstruct them. This enables the model to learn more generalized features from unlabeled data. We evaluate our proposed framework with three independent sleep datasets, which have shown significant improvement compared to supervised learning models.

Self-supervised Meta-learning for the Application of Federated Learning on the Medical Domain (연합학습의 의료분야 적용을 위한 자기지도 메타러닝)

  • Kong, Heesan;Kim, Kwangsu
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.4
    • /
    • pp.27-40
    • /
    • 2022
  • Medical AI, which has lately made significant advances, is playing a vital role, such as assisting clinicians with diagnosis and decision-making. The field of chest X-rays, in particular, is attracting a lot of attention since it is important for accessibility and identification of chest diseases, as well as the current COVID-19 pandemic. However, despite the vast amount of data, there remains a limit to developing an effective AI model due to a lack of labeled data. A research that used federated learning on chest X-ray data to lessen this difficulty has emerged, although it still has the following limitations. 1) It does not consider the problems that may occur in the Non-IID environment. 2) Even in the federated learning environment, there is still a shortage of labeled data of clients. We propose a method to solve the above problems by using the self-supervised learning model as a global model of federated learning. To that aim, we investigate a self-supervised learning methods suited for federated learning using chest X-ray data and demonstrate the benefits of adopting the self-supervised learning model for federated learning.

Facial Expression Recognition through Self-supervised Learning for Predicting Face Image Sequence

  • Yoon, Yeo-Chan;Kim, Soo Kyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.9
    • /
    • pp.41-47
    • /
    • 2022
  • In this paper, we propose a new and simple self-supervised learning method that predicts the middle image of a face image sequence for automatic expression recognition. Automatic facial expression recognition can achieve high performance through deep learning methods, however, generally requires a expensive large data set. The size of the data set and the performance of the algorithm are tend to be proportional. The proposed method learns latent deep representation of a face through self-supervised learning using an existing dataset without constructing an additional dataset. Then it transfers the learned parameter to new facial expression reorganization model for improving the performance of automatic expression recognition. The proposed method showed high performance improvement for two datasets, CK+ and AFEW 8.0, and showed that the proposed method can achieve a great effect.

A Study on Pattern Recognition with Self-Organized Supervised Learning (자기조직화 교사 학습에 의한 패턴인식에 관한 연구)

  • Park, Chan-Ho
    • The Journal of Information Technology
    • /
    • v.5 no.2
    • /
    • pp.17-26
    • /
    • 2002
  • On this paper, we propose SOSL(Self-Organized Supervised Learning) and it's architecture SOSL is hybrid type neural network. It consists of several CBP (Component Back Propagation) neural networks, and a modified PCA neural networks. CBP neural networks perform supervised learning procedure in parallel to clustered and complex input patterns. Modified PCA networks perform it's learning in order to transform dimensions of original input patterns to lower dimensions by clustering and local projection. Proposed SOSL can effectively apply to neural network learning with large input patterns results in huge networks size.

  • PDF

A Study on Enhanced Self-Generation Supervised Learning Algorithm for Image Recognition (영상 인식을 위한 개선된 자가 생성 지도 학습 알고리듬에 관한 연구)

  • Kim, Tae-Kyung;Kim, Kwang-Baek;Paik, Joon-Ki
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.2C
    • /
    • pp.31-40
    • /
    • 2005
  • we propose an enhanced self-generation supervised algorithm that by combining an ART algorithm and the delta-bar-delta method. Form the input layer to the hidden layer, ART-1 and ART-2 are used to produce nodes, respectively. A winner-take-all method is adopted to the connection weight adaption so that a stored pattern for some pattern is updated. we test the recognition of student identification, a certificate of residence, and an identifier from container that require nodes of hidden layers in neural network. In simulation results, the proposed self-generation supervised learning algorithm reduces the possibility of local minima and improves learning speed and paralysis than conventional neural networks.

Fuzzy Supervised Learning Algorithm by using Self-generation (Self-generation을 이용한 퍼지 지도 학습 알고리즘)

  • 김광백
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.7
    • /
    • pp.1312-1320
    • /
    • 2003
  • In this paper, we consider a multilayer neural network, with a single hidden layer. Error backpropagation learning method used widely in multilayer neural networks has a possibility of local minima due to the inadequate weights and the insufficient number of hidden nodes. So we propose a fuzzy supervised learning algorithm by using self-generation that self-generates hidden nodes by the compound fuzzy single layer perceptron and modified ART1. From the input layer to hidden layer, a modified ART1 is used to produce nodes. And winner take-all method is adopted to the connection weight adaptation, so that a stored pattern for some pattern gets updated. The proposed method has applied to the student identification card images. In simulation results, the proposed method reduces a possibility of local minima and improves learning speed and paralysis than the conventional error backpropagation learning algorithm.

  • PDF