• Title/Summary/Keyword: Few-Shot learning

Search Result 40, Processing Time 0.034 seconds

Intra-class Local Descriptor-based Prototypical Network for Few-Shot Learning

  • Huang, Xi-Lang;Choi, Seon Han
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.1
    • /
    • pp.52-60
    • /
    • 2022
  • Few-shot learning is a sub-area of machine learning problems, which aims to classify target images that only contain a few labeled samples for training. As a representative few-shot learning method, the Prototypical network has been received much attention due to its simplicity and promising results. However, the Prototypical network uses the sample mean of samples from the same class as the prototypes of that class, which easily results in learning uncharacteristic features in the low-data scenery. In this study, we propose to use local descriptors (i.e., patches along the channel within feature maps) from the same class to explicitly obtain more representative prototypes for Prototypical Network so that significant intra-class feature information can be maintained and thus improving the classification performance on few-shot learning tasks. Experimental results on various benchmark datasets including mini-ImageNet, CUB-200-2011, and tiered-ImageNet show that the proposed method can learn more discriminative intra-class features by the local descriptors and obtain more generic prototype representations under the few-shot setting.

Few-shot learning using the median prototype of the support set (Support set의 중앙값 prototype을 활용한 few-shot 학습)

  • Eu Tteum Baek
    • Smart Media Journal
    • /
    • v.12 no.1
    • /
    • pp.24-31
    • /
    • 2023
  • Meta-learning is metacognition that instantly distinguishes between knowing and unknown. It is a learning method that adapts and solves new problems by self-learning with a small amount of data.A few-shot learning method is a type of meta-learning method that accurately predicts query data even with a very small support set. In this study, we propose a method to solve the limitations of the prototype created with the mean-point vector of each class. For this purpose, we use the few-shot learning method that created the prototype used in the few-shot learning method as the median prototype. For quantitative evaluation, a handwriting recognition dataset and mini-Imagenet dataset were used and compared with the existing method. Through the experimental results, it was confirmed that the performance was improved compared to the existing method.

Recent advances in few-shot learning for image domain: a survey (이미지 분석을 위한 퓨샷 학습의 최신 연구동향)

  • Ho-Sik Seok
    • Journal of IKEEE
    • /
    • v.27 no.4
    • /
    • pp.537-547
    • /
    • 2023
  • In many domains, lack of data inhibits adoption of advanced machine learning models. Recently, Few-Shot Learning (FSL) has been actively studied to tackle this problem. Utilizing prior knowledge obtained through observations on related domains, FSL achieved significant performance with only a few samples. In this paper, we present a survey on FSL in terms of data augmentation, embedding and metric learning, and meta-learning. In addition to interesting researches, we also introduce major benchmark datasets. FSL is widely adopted in various domains, but we focus on image analysis in this paper.

A Nested Named Entity Recognition Model Robust in Few-shot Learning Environments using Label Information (라벨 정보를 이용한 Few-shot Learning 환경에 강건한 중첩 개체명 인식 모델)

  • Hyunsun Hwang;Changki Lee;Wooyoung Go;Myungchul Kang
    • Annual Conference on Human and Language Technology
    • /
    • 2023.10a
    • /
    • pp.622-626
    • /
    • 2023
  • 중첩 개체명 인식(Nested Named Entity Recognition)은 하나의 개체명 표현 안에 다른 개체명 표현이 들어 있는 중첩 구조의 개체명을 인식하는 작업으로, 중첩 개체명 인식을 위한 학습데이터 구축 작업은 일반 개체명 인식 학습데이터 구축보다 어렵다는 문제가 있다. 본 논문에서는 이러한 문제를 해결하기 위해 Few-shot Learning 환경에 강건한 중첩 개체명 인식 모델을 제안한다. 이를 위해, 기존의 Biaffine 중첩 개체명 인식 모델의 출력 레이어를 라벨 의미 정보를 활용하도록 변경하여 학습데이터가 적은 환경에서 중첩 개체명 인식의 성능을 향상시키도록 하였다. 실험 결과 GENIA 중첩 개체명 인식 데이터의 5-shot, 10-shot, 20-shot 환경에서 기존의 Biaffine 모델보다 평균 10%p이상의 높은 F1-measure 성능을 보였다.

  • PDF

Prompt-based Full-Shot and Few-Shot Learning for Diagnosing Dementia and Schizophrenia (Prompt 기반의 Full-Shot Learning과 Few-Shot Learning을 이용한 알츠하이머병 치매와 조현병 진단)

  • Min-Kyo Jung;Seung-Hoon Na;Ko Woon Kim;Byoung-Soo Shin;Young-Chul Chung
    • Annual Conference on Human and Language Technology
    • /
    • 2022.10a
    • /
    • pp.47-52
    • /
    • 2022
  • 환자와 주변인들에게 다양한 문제를 야기하는 치매와 조현병 진단을 위한 모델을 제안한다. 치매와 조현병 진단을 위해 프로토콜에 따라 녹음한 의사와 내담자 음성 시료를 전사 작업하여 분류 태스크를 수행하였다. 사전 학습한 언어 모델의 MLM Head를 이용해 분류 태스크를 수행하는 Prompt 기반의 분류 모델을 제안하였다. 또한 많은 수의 데이터 수를 확보하기 어려운 의료 분야에 효율적인 Few-Shot 학습 방식을 이용하였다. CLS 토큰을 미세조정하는 일반적 학습 방식의 Baseline과 비교해 Full-Shot 실험에서 7개 태스크 중 1개 태스크에서 macro, micro-F1 점수 모두 향상되었고, 3개 태스크에서 하나의 F1 점수만 향샹된 것을 확인 하였다. 반면, Few-Shot 실험에서는 7개 태스크 중 2개 태스크에서 macro, micro-F1 점수가 모두 향상되었고, 2개 태스크에서 하나의 F1 점수만 향상되었다.

  • PDF

Extrapolation of wind pressure for low-rise buildings at different scales using few-shot learning

  • Yanmo Weng;Stephanie G. Paal
    • Wind and Structures
    • /
    • v.36 no.6
    • /
    • pp.367-377
    • /
    • 2023
  • This study proposes a few-shot learning model for extrapolating the wind pressure of scaled experiments to full-scale measurements. The proposed ML model can use scaled experimental data and a few full-scale tests to accurately predict the remaining full-scale data points (for new specimens). This model focuses on extrapolating the prediction to different scales while existing approaches are not capable of accurately extrapolating from scaled data to full-scale data in the wind engineering domain. Also, the scaling issue observed in wind tunnel tests can be partially resolved via the proposed approach. The proposed model obtained a low mean-squared error and a high coefficient of determination for the mean and standard deviation wind pressure coefficients of the full-scale dataset. A parametric study is carried out to investigate the influence of the number of selected shots. This technique is the first of its kind as it is the first time an ML model has been used in the wind engineering field to deal with extrapolation in wind performance prediction. With the advantages of the few-shot learning model, physical wind tunnel experiments can be reduced to a great extent. The few-shot learning model yields a robust, efficient, and accurate alternative to extrapolating the prediction performance of structures from various model scales to full-scale.

Image Classification Using Few-Shot Learning (Few-Shot 학습을 이용한 영상 분류)

  • Shin, Seong-Yoon;Lee, Hyun-Chang;Shin, Kwang-Seong
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.07a
    • /
    • pp.681-682
    • /
    • 2021
  • 본 논문에서는 소규모 데이터 셋의 이미지 분류 작업에서 모델 과적 합 및 비 수렴을 해결하고 분류의 정확도를 높이는 데 주로 사용되는 few-shot 학습을 기반으로 한 새로운 이미지 분류 방법을 제안합니다.

  • PDF

Learning Deep Representation by Increasing ConvNets Depth for Few Shot Learning

  • Fabian, H.S. Tan;Kang, Dae-Ki
    • International journal of advanced smart convergence
    • /
    • v.8 no.4
    • /
    • pp.75-81
    • /
    • 2019
  • Though recent advancement of deep learning methods have provided satisfactory results from large data domain, somehow yield poor performance on few-shot classification tasks. In order to train a model with strong performance, i.e. deep convolutional neural network, it depends heavily on huge dataset and the labeled classes of the dataset can be extremely humongous. The cost of human annotation and scarcity of the data among the classes have drastically limited the capability of current image classification model. On the contrary, humans are excellent in terms of learning or recognizing new unseen classes with merely small set of labeled examples. Few-shot learning aims to train a classification model with limited labeled samples to recognize new classes that have neverseen during training process. In this paper, we increase the backbone depth of the embedding network in orderto learn the variation between the intra-class. By increasing the network depth of the embedding module, we are able to achieve competitive performance due to the minimized intra-class variation.

Few-shot Aerial Image Segmentation with Mask-Guided Attention (마스크-보조 어텐션 기법을 활용한 항공 영상에서의 퓨-샷 의미론적 분할)

  • Kwon, Hyeongjun;Song, Taeyong;Lee, Tae-Young;Ahn, Jongsik;Sohn, Kwanghoon
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.5
    • /
    • pp.685-694
    • /
    • 2022
  • The goal of few-shot semantic segmentation is to build a network that quickly adapts to novel classes with extreme data shortage regimes. Most existing few-shot segmentation methods leverage single or multiple prototypes from extracted support features. Although there have been promising results for natural images, these methods are not directly applicable to the aerial image domain. A key factor in few-shot segmentation on aerial images is to effectively exploit information that is robust against extreme changes in background and object scales. In this paper, we propose a Mask-Guided Attention module to extract more comprehensive support features for few-shot segmentation in aerial images. Taking advantage of the support ground-truth masks, the area correlated to the foreground object is highlighted and enables the support encoder to extract comprehensive support features with contextual information. To facilitate reproducible studies of the task of few-shot semantic segmentation in aerial images, we further present the few-shot segmentation benchmark iSAID-, which is constructed from a large-scale iSAID dataset. Extensive experimental results including comparisons with the state-of-the-art methods and ablation studies demonstrate the effectiveness of the proposed method.

Improving Few-Shot Learning through Self-Distillation (Self-Distillation을 활용한 Few-Shot 학습 개선)

  • Kim, Tae-Hun;Choo, Jae-Gul
    • Annual Conference of KIPS
    • /
    • 2018.10a
    • /
    • pp.617-620
    • /
    • 2018
  • 딥러닝 기술에 있어서 대량의 학습 데이터가 필요하다는 한계점을 극복하기 위한 시도로서, 적은 데이터 만으로도 좋은 성능을 낼 수 있는 few-shot 학습 모델이 꾸준히 발전하고 있다. 하지만 few-shot 학습 모델의 가장 큰 단점인 적은 데이터로 인한 과적합 문제는 여전히 어려운 숙제로 남아있다. 본 논문에서는 모델 압축에 사용되는 distillation 기법을 사용하여 few-shot 학습 모델의 학습 문제를 개선하고자 한다. 이를 위해 대표적인 few-shot 모델인 Siamese Networks, Prototypical Networks, Matching Networks에 각각 distillation을 적용하였다. 본 논문의 실험결과로써 단순히 결과값에 대한 참/거짓 뿐만 아니라, 참/거짓에 대한 신뢰도까지 같이 학습함으로써 few-shot 모델의 학습 문제 개선에 도움이 된다는 것을 실험적으로 증명하였다.