• Title/Summary/Keyword: Self-supervised learning

Search Result 97, Processing Time 0.031 seconds

Attentive Transfer Learning via Self-supervised Learning for Cervical Dysplasia Diagnosis

  • Chae, Jinyeong;Zimmermann, Roger;Kim, Dongho;Kim, Jihie
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.453-461
    • /
    • 2021
  • Many deep learning approaches have been studied for image classification in computer vision. However, there are not enough data to generate accurate models in medical fields, and many datasets are not annotated. This study presents a new method that can use both unlabeled and labeled data. The proposed method is applied to classify cervix images into normal versus cancerous, and we demonstrate the results. First, we use a patch self-supervised learning for training the global context of the image using an unlabeled image dataset. Second, we generate a classifier model by using the transferred knowledge from self-supervised learning. We also apply attention learning to capture the local features of the image. The combined method provides better performance than state-of-the-art approaches in accuracy and sensitivity.

Software Fault Prediction using Semi-supervised Learning Methods (세미감독형 학습 기법을 사용한 소프트웨어 결함 예측)

  • Hong, Euyseok
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.19 no.3
    • /
    • pp.127-133
    • /
    • 2019
  • Most studies of software fault prediction have been about supervised learning models that use only labeled training data. Although supervised learning usually shows high prediction performance, most development groups do not have sufficient labeled data. Unsupervised learning models that use only unlabeled data for training are difficult to build and show poor performance. Semi-supervised learning models that use both labeled data and unlabeled data can solve these problems. Self-training technique requires the fewest assumptions and constraints among semi-supervised techniques. In this paper, we implemented several models using self-training algorithms and evaluated them using Accuracy and AUC. As a result, YATSI showed the best performance.

Pretext Task Analysis for Self-Supervised Learning Application of Medical Data (의료 데이터의 자기지도학습 적용을 위한 pretext task 분석)

  • Kong, Heesan;Park, Jaehun;Kim, Kwangsu
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.38-40
    • /
    • 2021
  • Medical domain has a massive number of data records without the response value. Self-supervised learning is a suitable method for medical data since it learns pretext-task and supervision, which the model can understand the semantic representation of data without response values. However, since self-supervised learning performance depends on the expression learned by the pretext-task, it is necessary to define an appropriate Pretext-task with data feature consideration. In this paper, to actively exploit the unlabeled medical data into artificial intelligence research, experimentally find pretext-tasks that suitable for the medical data and analyze the result. We use the x-ray image dataset which is effectively utilizable for the medical domain.

  • PDF

Multimodal Supervised Contrastive Learning for Crop Disease Diagnosis (멀티 모달 지도 대조 학습을 이용한 농작물 병해 진단 예측 방법)

  • Hyunseok Lee;Doyeob Yeo;Gyu-Sung Ham;Kanghan Oh
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.6
    • /
    • pp.285-292
    • /
    • 2023
  • With the wide spread of smart farms and the advancements in IoT technology, it is easy to obtain additional data in addition to crop images. Consequently, deep learning-based crop disease diagnosis research utilizing multimodal data has become important. This study proposes a crop disease diagnosis method using multimodal supervised contrastive learning by expanding upon the multimodal self-supervised learning. RandAugment method was used to augment crop image and time series of environment data. These augmented data passed through encoder and projection head for each modality, yielding low-dimensional features. Subsequently, the proposed multimodal supervised contrastive loss helped features from the same class get closer while pushing apart those from different classes. Following this, the pretrained model was fine-tuned for crop disease diagnosis. The visualization of t-SNE result and comparative assessments of crop disease diagnosis performance substantiate that the proposed method has superior performance than multimodal self-supervised learning.

Self-supervised Meta-learning for the Application of Federated Learning on the Medical Domain (연합학습의 의료분야 적용을 위한 자기지도 메타러닝)

  • Kong, Heesan;Kim, Kwangsu
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.4
    • /
    • pp.27-40
    • /
    • 2022
  • Medical AI, which has lately made significant advances, is playing a vital role, such as assisting clinicians with diagnosis and decision-making. The field of chest X-rays, in particular, is attracting a lot of attention since it is important for accessibility and identification of chest diseases, as well as the current COVID-19 pandemic. However, despite the vast amount of data, there remains a limit to developing an effective AI model due to a lack of labeled data. A research that used federated learning on chest X-ray data to lessen this difficulty has emerged, although it still has the following limitations. 1) It does not consider the problems that may occur in the Non-IID environment. 2) Even in the federated learning environment, there is still a shortage of labeled data of clients. We propose a method to solve the above problems by using the self-supervised learning model as a global model of federated learning. To that aim, we investigate a self-supervised learning methods suited for federated learning using chest X-ray data and demonstrate the benefits of adopting the self-supervised learning model for federated learning.

Facial Expression Recognition through Self-supervised Learning for Predicting Face Image Sequence

  • Yoon, Yeo-Chan;Kim, Soo Kyun
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.9
    • /
    • pp.41-47
    • /
    • 2022
  • In this paper, we propose a new and simple self-supervised learning method that predicts the middle image of a face image sequence for automatic expression recognition. Automatic facial expression recognition can achieve high performance through deep learning methods, however, generally requires a expensive large data set. The size of the data set and the performance of the algorithm are tend to be proportional. The proposed method learns latent deep representation of a face through self-supervised learning using an existing dataset without constructing an additional dataset. Then it transfers the learned parameter to new facial expression reorganization model for improving the performance of automatic expression recognition. The proposed method showed high performance improvement for two datasets, CK+ and AFEW 8.0, and showed that the proposed method can achieve a great effect.

A Study on Pattern Recognition with Self-Organized Supervised Learning (자기조직화 교사 학습에 의한 패턴인식에 관한 연구)

  • Park, Chan-Ho
    • The Journal of Information Technology
    • /
    • v.5 no.2
    • /
    • pp.17-26
    • /
    • 2002
  • On this paper, we propose SOSL(Self-Organized Supervised Learning) and it's architecture SOSL is hybrid type neural network. It consists of several CBP (Component Back Propagation) neural networks, and a modified PCA neural networks. CBP neural networks perform supervised learning procedure in parallel to clustered and complex input patterns. Modified PCA networks perform it's learning in order to transform dimensions of original input patterns to lower dimensions by clustering and local projection. Proposed SOSL can effectively apply to neural network learning with large input patterns results in huge networks size.

  • PDF

A Study on Enhanced Self-Generation Supervised Learning Algorithm for Image Recognition (영상 인식을 위한 개선된 자가 생성 지도 학습 알고리듬에 관한 연구)

  • Kim, Tae-Kyung;Kim, Kwang-Baek;Paik, Joon-Ki
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.30 no.2C
    • /
    • pp.31-40
    • /
    • 2005
  • we propose an enhanced self-generation supervised algorithm that by combining an ART algorithm and the delta-bar-delta method. Form the input layer to the hidden layer, ART-1 and ART-2 are used to produce nodes, respectively. A winner-take-all method is adopted to the connection weight adaption so that a stored pattern for some pattern is updated. we test the recognition of student identification, a certificate of residence, and an identifier from container that require nodes of hidden layers in neural network. In simulation results, the proposed self-generation supervised learning algorithm reduces the possibility of local minima and improves learning speed and paralysis than conventional neural networks.

Fuzzy Supervised Learning Algorithm by using Self-generation (Self-generation을 이용한 퍼지 지도 학습 알고리즘)

  • 김광백
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.7
    • /
    • pp.1312-1320
    • /
    • 2003
  • In this paper, we consider a multilayer neural network, with a single hidden layer. Error backpropagation learning method used widely in multilayer neural networks has a possibility of local minima due to the inadequate weights and the insufficient number of hidden nodes. So we propose a fuzzy supervised learning algorithm by using self-generation that self-generates hidden nodes by the compound fuzzy single layer perceptron and modified ART1. From the input layer to hidden layer, a modified ART1 is used to produce nodes. And winner take-all method is adopted to the connection weight adaptation, so that a stored pattern for some pattern gets updated. The proposed method has applied to the student identification card images. In simulation results, the proposed method reduces a possibility of local minima and improves learning speed and paralysis than the conventional error backpropagation learning algorithm.

  • PDF

An Analysis of the methods to alleviate the cost of data labeling in Deep learning (딥 러닝에서 Labeling 부담을 줄이기 위한 연구분석)

  • Han, Seokmin
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.1
    • /
    • pp.545-550
    • /
    • 2022
  • In Deep Learning method, it is well known that it requires large amount of data to train the deep neural network. And it also requires the labeling of each data to fully train the neural network, which means that experts should spend lots of time to provide the labeling. To alleviate the problem of time-consuming labeling process, some methods have been suggested such as weak-supervised method, one-shot learning, self-supervised, suggestive learning, and so on. In this manuscript, those methods are analyzed and its possible future direction of the research is suggested.