• Title/Summary/Keyword: Language Models

Search Result 872, Processing Time 0.025 seconds

Zero-shot Korean Sentiment Analysis with Large Language Models: Comparison with Pre-trained Language Models

  • Soon-Chan Kwon;Dong-Hee Lee;Beak-Cheol Jang
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.2
    • /
    • pp.43-50
    • /
    • 2024
  • This paper evaluates the Korean sentiment analysis performance of large language models like GPT-3.5 and GPT-4 using a zero-shot approach facilitated by the ChatGPT API, comparing them to pre-trained Korean models such as KoBERT. Through experiments utilizing various Korean sentiment analysis datasets in fields like movies, gaming, and shopping, the efficiency of these models is validated. The results reveal that the LMKor-ELECTRA model displayed the highest performance based on F1-score, while GPT-4 particularly achieved high accuracy and F1-scores in movie and shopping datasets. This indicates that large language models can perform effectively in Korean sentiment analysis without prior training on specific datasets, suggesting their potential in zero-shot learning. However, relatively lower performance in some datasets highlights the limitations of the zero-shot based methodology. This study explores the feasibility of using large language models for Korean sentiment analysis, providing significant implications for future research in this area.

Recent R&D Trends for Pretrained Language Model (딥러닝 사전학습 언어모델 기술 동향)

  • Lim, J.H.;Kim, H.K.;Kim, Y.K.
    • Electronics and Telecommunications Trends
    • /
    • v.35 no.3
    • /
    • pp.9-19
    • /
    • 2020
  • Recently, a technique for applying a deep learning language model pretrained from a large corpus to fine-tuning for each application task has been widely used as a language processing technology. The pretrained language model shows higher performance and satisfactory generalization performance than existing methods. This paper introduces the major research trends related to deep learning pretrained language models in the field of language processing. We describe in detail the motivations, models, learning methods, and results of the BERT language model that had significant influence on subsequent studies. Subsequently, we introduce the results of language model studies after BERT, focusing on SpanBERT, RoBERTa, ALBERT, BART, and ELECTRA. Finally, we introduce the KorBERT pretrained language model, which shows satisfactory performance in Korean language. In addition, we introduce techniques on how to apply the pretrained language model to Korean (agglutinative) language, which consists of a combination of content and functional morphemes, unlike English (refractive) language whose endings change depending on the application.

Software Model Integration Using Metadata Model Based on Linked Data (Linked Data 기반의 메타데이타 모델을 활용한 소프트웨어 모델 통합)

  • Kim, Dae-Hwan;Jeong, Chan-Ki
    • Journal of Information Technology Services
    • /
    • v.12 no.3
    • /
    • pp.311-321
    • /
    • 2013
  • In the community of software engineering, diverse modeling languages are used for representing all relevant information in the form of models. Also many different models such as business model, business process model, product models, interface models etc. are generated through software life cycles. In this situation, models need to be integrated for enterprise integration and enhancement of software productivity. Researchers propose rebuilding models by a specific modeling language, using a intemediate modeling language and using common reference for model integration. However, in the current approach it requires a lot of cost and time to integrate models. Also it is difficult to identify common objects from several models and to update objects in the repository of common model objects. This paper proposes software model integration using metadata model based on Linked data. We verify the effectiveness of the proposed approach through a case study.

Language Modeling Approaches to Information Retrieval

  • Banerjee, Protima;Han, Hyo-Il
    • Journal of Computing Science and Engineering
    • /
    • v.3 no.3
    • /
    • pp.143-164
    • /
    • 2009
  • This article surveys recent research in the area of language modeling (sometimes called statistical language modeling) approaches to information retrieval. Language modeling is a formal probabilistic retrieval framework with roots in speech recognition and natural language processing. The underlying assumption of language modeling is that human language generation is a random process; the goal is to model that process via a generative statistical model. In this article, we discuss current research in the application of language modeling to information retrieval, the role of semantics in the language modeling framework, cluster-based language models, use of language modeling for XML retrieval and future trends.

Updated Primer on Generative Artificial Intelligence and Large Language Models in Medical Imaging for Medical Professionals

  • Kiduk Kim;Kyungjin Cho;Ryoungwoo Jang;Sunggu Kyung;Soyoung Lee;Sungwon Ham;Edward Choi;Gil-Sun Hong;Namkug Kim
    • Korean Journal of Radiology
    • /
    • v.25 no.3
    • /
    • pp.224-242
    • /
    • 2024
  • The emergence of Chat Generative Pre-trained Transformer (ChatGPT), a chatbot developed by OpenAI, has garnered interest in the application of generative artificial intelligence (AI) models in the medical field. This review summarizes different generative AI models and their potential applications in the field of medicine and explores the evolving landscape of Generative Adversarial Networks and diffusion models since the introduction of generative AI models. These models have made valuable contributions to the field of radiology. Furthermore, this review also explores the significance of synthetic data in addressing privacy concerns and augmenting data diversity and quality within the medical domain, in addition to emphasizing the role of inversion in the investigation of generative models and outlining an approach to replicate this process. We provide an overview of Large Language Models, such as GPTs and bidirectional encoder representations (BERTs), that focus on prominent representatives and discuss recent initiatives involving language-vision models in radiology, including innovative large language and vision assistant for biomedicine (LLaVa-Med), to illustrate their practical application. This comprehensive review offers insights into the wide-ranging applications of generative AI models in clinical research and emphasizes their transformative potential.

Performance Evaluation of Pre-trained Language Models in Multi-Goal Conversational Recommender Systems (다중목표 대화형 추천시스템을 위한 사전 학습된 언어모델들에 대한 성능 평가)

  • Taeho Kim;Hyung-Jun Jang;Sang-Wook Kim
    • Smart Media Journal
    • /
    • v.12 no.6
    • /
    • pp.35-40
    • /
    • 2023
  • In this study paper, we examine pre-trained language models used in Multi-Goal Conversational Recommender Systems (MG-CRS), comparing and analyzing their performances of various pre-trained language models. Specifically, we investigates the impact of the sizes of language models on the performance of MG-CRS. The study targets three types of language models - of BERT, GPT2, and BART, and measures and compares their accuracy in two tasks of 'type prediction' and 'topic prediction' on the MG-CRS dataset, DuRecDial 2.0. Experimental results show that all models demonstrated excellent performance in the type prediction task, but there were notable provide significant performance differences in performance depending on among the models or based on their sizes in the topic prediction task. Based on these findings, the study provides directions for improving the performance of MG-CRS.

Probing Sentence Embeddings in L2 Learners' LSTM Neural Language Models Using Adaptation Learning

  • Kim, Euhee
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.3
    • /
    • pp.13-23
    • /
    • 2022
  • In this study we leveraged a probing method to evaluate how a pre-trained L2 LSTM language model represents sentences with relative and coordinate clauses. The probing experiment employed adapted models based on the pre-trained L2 language models to trace the syntactic properties of sentence embedding vector representations. The dataset for probing was automatically generated using several templates related to different sentence structures. To classify the syntactic properties of sentences for each probing task, we measured the adaptation effects of the language models using syntactic priming. We performed linear mixed-effects model analyses to analyze the relation between adaptation effects in a complex statistical manner and reveal how the L2 language models represent syntactic features for English sentences. When the L2 language models were compared with the baseline L1 Gulordava language models, the analogous results were found for each probing task. In addition, it was confirmed that the L2 language models contain syntactic features of relative and coordinate clauses hierarchically in the sentence embedding representations.

Analysis of Discriminatory Patterns in Performing Arts Recognized by Large Language Models (LLMs): Focused on ChatGPT (거대언어모델(LLM)이 인식하는 공연예술의 차별 양상 분석: ChatGPT를 중심으로)

  • Jiae Choi
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.401-418
    • /
    • 2023
  • Recently, the socio-economic interest in Large Language Models (LLMs) has been growing due to the emergence of ChatGPT. As a type of generative AI, LLMs have reached the level of script creation. In this regard, it is important to address the issue of discrimination (sexism, racism, religious discrimination, ageism, etc.) in the performing arts in general or in specific performing arts works or organizations in a large language model that will be widely used by the general public and professionals. However, there has not yet been a full-scale investigation and discussion on the issue of discrimination in the performing arts in large-scale language models. Therefore, the purpose of this study is to textually analyze the perceptions of discrimination issues in the performing arts from LMMs and to derive implications for the performing arts field and the development of LMMs. First, BBQ (Bias Benchmark for QA) questions and measures for nine discrimination issues were used to measure the sensitivity to discrimination of the giant language models, and the answers derived from the representative giant language models were verified by performing arts experts to see if there were any parts of the giant language models' misperceptions, and then the giant language models' perceptions of the ethics of discriminatory views in the performing arts field were analyzed through the content analysis method. As a result of the analysis, implications for the performing arts field and points to be noted in the development of large-scale linguistic models were derived and discussed.

XBRL-Based Representation and Sharing of Decision Models (XBRL 기반의 의사결정 모형 표현과 공유)

  • Kim, Hyoung-Do;Park, Chan-Kwon;Yum, Ji-Hwan;Lee, Sung-Hoon
    • Journal of Information Technology Applications and Management
    • /
    • v.14 no.2
    • /
    • pp.117-127
    • /
    • 2007
  • Using an exchange standard, we can design an open architecture for the interchange of decision models and data. XML (eXtensible Markup Language) provides a general framework for creating such a standard. Although XML -based model representation languages such as OOSML were proposed, they are partly limited in expression capability, flexibility, generality, etc. This paper proposes a new method for expressing and sharing decision models and data based on XBRL (eXtensible Business Reporting Language), which is a XML language specialized in business reporting. We have developed a XBRL taxonomy for decision models with the concepts and relationships of a representative modeling framework, SM (Structured Modeling). The method allows for expressing data as well as decision models in a consistent and flexible manner. Diverse dependencies between components of SM models can also be affluently expressed.

  • PDF

Dependency Structure Applied to Language Modeling for Information Retrieval

  • Lee, Chang-Ki;Lee, Gary Geun-Bae;Jang, Myung-Gil
    • ETRI Journal
    • /
    • v.28 no.3
    • /
    • pp.337-346
    • /
    • 2006
  • In this paper, we propose a new language model, namely, a dependency structure language model, for information retrieval to compensate for the weaknesses of unigram and bigram language models. The dependency structure language model is based on the first-order dependency model and the dependency parse tree generated by a linguistic parser. So, long-distance dependencies can be naturally captured by the dependency structure language model. We carried out extensive experiments to verify the proposed model, where the dependency structure model gives a better performance than recently proposed language models and the Okapi BM25 method, and the dependency structure is more effective than unigram and bigram in language modeling for information retrieval.

  • PDF