• Title/Summary/Keyword: Language generation

Search Result 733, Processing Time 0.032 seconds

The Influence of Attitudes toward Korean Language and Motivational Intensity on Korean Proficiency of Korean Residents in Japan (재일 동포의 한국어에 대한 태도와 학습 동기 강도가 한국어 능력에 미치는 영향)

  • Kim, Heesang;Kim, Hyoeun
    • Journal of Korean language education
    • /
    • v.28 no.1
    • /
    • pp.49-78
    • /
    • 2017
  • This study aims to analyze the effect of attitudes of Korean residents in Japan towards learning the Korean language and their motivational intensity on their Korean proficiency. Data for this study came from a survey on language use of Korean residents in Japan which was conducted in 2016, and questionnaire items referred to language attitude, language use and the degree of understanding language; language use; language learning and Korean ethnic identity. The main results are as follows. First, there were significant differences in Korean language proficiency depending on age, education levels and generation. Second, the control for socio-demographic characteristics, the influence of attitudes towards Korean language on Korean proficiency was statistically significant. However, Korean proficiency was not significantly influenced by motivational intensity. Lastly, moderated effects of immigrant generation in the relation between Korean language attitudes and Korean proficiency were significant. Therefore, the effect of Korean language attitudes on Korean proficiency was more influential on second and third generation Korean-Japanese learners than first generation Korean-Japanese learners. Based on these results, this study suggests that in order to promote Korean language education for Korean residents in Japan, it is required to build positive attitudes toward Korean language, and to consider immigrant generation as a major factor.

Subword Neural Language Generation with Unlikelihood Training

  • Iqbal, Salahuddin Muhammad;Kang, Dae-Ki
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.12 no.2
    • /
    • pp.45-50
    • /
    • 2020
  • A Language model with neural networks commonly trained with likelihood loss. Such that the model can learn the sequence of human text. State-of-the-art results achieved in various language generation tasks, e.g., text summarization, dialogue response generation, and text generation, by utilizing the language model's next token output probabilities. Monotonous and boring outputs are a well-known problem of this model, yet only a few solutions proposed to address this problem. Several decoding techniques proposed to suppress repetitive tokens. Unlikelihood training approached this problem by penalizing candidate tokens probabilities if the tokens already seen in previous steps. While the method successfully showed a less repetitive generated token, the method has a large memory consumption because of the training need a big vocabulary size. We effectively reduced memory footprint by encoding words as sequences of subword units. Finally, we report competitive results with token level unlikelihood training in several automatic evaluations compared to the previous work.

A Frame-based Approach to Text Generation

  • Le, Huong Thanh
    • Proceedings of the Korean Society for Language and Information Conference
    • /
    • 2007.11a
    • /
    • pp.192-201
    • /
    • 2007
  • This paper is a study on constructing a natural language interface to database, concentrating on generating textual answers. TGEN, a system that generates textual answer from query result tables is presented. The TGEN architecture guarantees its portability across domains. A combination of a frame-based approach and natural language generation techniques in the TGEN provides text fluency and text flexibility. The implementation result shows that this approach is feasible while a deep NLG approach is still far to be reached.

  • PDF

Reproducibility of Hemispheric Language Dominance by Noun, Verb, Adjective and Adverb Generation Paradigms in Functional Magnetic Resonance Imaging of Normal Volunteers (정상성인의 뇌기능적 자기공명영상에서 명사, 동사, 형용사 그리고 부사 만들기 과제들에 대한 언어영역편재화의 재현성에 관한 연구)

  • In Chan Song;Kee Hyun Chang;Chun Kee Chung;Sang Hyun Lee;Moon Hee Han
    • Investigative Magnetic Resonance Imaging
    • /
    • v.5 no.1
    • /
    • pp.24-32
    • /
    • 2001
  • Purpose : We investigated the reproducibility of language lateralization by 4 different word generation paradigms or the rest contents in each paradigm using functional magnetic resonance imaging in normal volunteers Materials and Methods Nine normal volunteers with left-handedness (mean age: 25 yrs) were examined on a 1.57 MR unit using a single-shot gradient echo epibold sequence. Four different word generation paradigms of noun, verb, adjective and adverb were used in each normal volunteer for investigating language system. In each paradigm, two different rest contents consisted of only seeing the " +" symbol or reading the meaningless letters. Each task consisted of 96 phases including 3 activations and 6 rests of 2 different contents. Two activation maps in one task were obtained under two different rest contents using the correlation method. We evaluated the detection rates of Broca and Wernicke areas and the differences of language lateralization among four different word generation paradigms, or between the rest contents. Results : The detection rates of Broca and Wernicke areas were over 67 % in 4 different language paradigms and there was no significant difference of them among language paradigms, or between two different rest contents. Language dominances, in all 4 different language paradigms, were shown to be consistent in 66 %, but were contrary with language paradigms in some subjects. The rest contents made no significant effect on dominant language dominance determination, but the success rates of the dominant language dominances determined from 4 language paradigms were higher in reading the meaningless letter (100%, n=9) than in only seeing "+" on screen at the rest task (78%, n=7).

  • PDF

A Survey of Automatic Code Generation from Natural Language

  • Shin, Jiho;Nam, Jaechang
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.537-555
    • /
    • 2021
  • Many researchers have carried out studies related to programming languages since the beginning of computer science. Besides programming with traditional programming languages (i.e., procedural, object-oriented, functional programming language, etc.), a new paradigm of programming is being carried out. It is programming with natural language. By programming with natural language, we expect that it will free our expressiveness in contrast to programming languages which have strong constraints in syntax. This paper surveys the approaches that generate source code automatically from a natural language description. We also categorize the approaches by their forms of input and output. Finally, we analyze the current trend of approaches and suggest the future direction of this research domain to improve automatic code generation with natural language. From the analysis, we state that researchers should work on customizing language models in the domain of source code and explore better representations of source code such as embedding techniques and pre-trained models which have been proved to work well on natural language processing tasks.

An Automatic Microcode Generation System Using a Microinstruction Description Language (마이크로명령어 기술언어를 사용한 마이크로코드 자동생성 시스템)

  • 이상정;조영훈;임인칠
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.28B no.7
    • /
    • pp.540-547
    • /
    • 1991
  • This paper proposes a machine in dependent automatic microcode generation system using a microtnstruction description language, MDL. The MDL, which has similar structure to C language, is a high-level microarchitecture description language. It defines the hardwaer elements and the operand selection of microoperartions. The proposed system generates microcode automatically by describing the structure information of a target microarchitectuer and accepting thebehavioral information of microoperations which are generated ad a intermediate language from HLML-C. This proposed system is implemented with C language and YACC on a SUN workstation (4.3 BSD).

  • PDF

Enhanced Regular Expression as a DGL for Generation of Synthetic Big Data

  • Kai, Cheng;Keisuke, Abe
    • Journal of Information Processing Systems
    • /
    • v.19 no.1
    • /
    • pp.1-16
    • /
    • 2023
  • Synthetic data generation is generally used in performance evaluation and function tests in data-intensive applications, as well as in various areas of data analytics, such as privacy-preserving data publishing (PPDP) and statistical disclosure limit/control. A significant amount of research has been conducted on tools and languages for data generation. However, existing tools and languages have been developed for specific purposes and are unsuitable for other domains. In this article, we propose a regular expression-based data generation language (DGL) for flexible big data generation. To achieve a general-purpose and powerful DGL, we enhanced the standard regular expressions to support the data domain, type/format inference, sequence and random generation, probability distributions, and resource reference. To efficiently implement the proposed language, we propose caching techniques for both the intermediate and database queries. We evaluated the proposed improvement experimentally.

A study on Implementation of English Sentence Generator using Lexical Functions (언어함수를 이용한 영문 생성기의 구현에 관한 연구)

  • 정희연;김희연;이웅재
    • Journal of Internet Computing and Services
    • /
    • v.1 no.2
    • /
    • pp.49-59
    • /
    • 2000
  • The majority of work done to date on natural language processing has focused on analysis and understanding of language, thus natural language generation had been relatively less attention than understanding, And people even tends to regard natural language generation CIS a simple reverse process of language understanding, However, need for natural language generation is growing rapidly as application systems, especially multi-language machine translation systems on the web, natural language interface systems, natural language query systems need more complex messages to generate, In this paper, we propose an algorithm to generate more flexible and natural sentence using lexical functions of Igor Mel'uk (Mel'uk & Zholkovsky, 1988) and systemic grammar.

  • PDF

Framework for evaluating code generation ability of large language models

  • Sangyeop Yeo;Yu-Seung Ma;Sang Cheol Kim;Hyungkook Jun;Taeho Kim
    • ETRI Journal
    • /
    • v.46 no.1
    • /
    • pp.106-117
    • /
    • 2024
  • Large language models (LLMs) have revolutionized various applications in natural language processing and exhibited proficiency in generating programming code. We propose a framework for evaluating the code generation ability of LLMs and introduce a new metric, pass-ratio@n, which captures the granularity of accuracy according to the pass rate of test cases. The framework is intended to be fully automatic to handle the repetitive work involved in generating prompts, conducting inferences, and executing the generated codes. A preliminary evaluation focusing on the prompt detail, problem publication date, and difficulty level demonstrates the successful integration of our framework with the LeetCode coding platform and highlights the applicability of the pass-ratio@n metric.

Design and Implementation of a Augmentative and Alternative Communication System Using Sentence Generation (문장생성에 의한 통신보조시스템의 설계 및 구현)

  • Woo Yo-Seop;Min Hong-Ki;Hwang Ein-Jeong
    • Journal of Korea Multimedia Society
    • /
    • v.8 no.9
    • /
    • pp.1248-1257
    • /
    • 2005
  • This paper designs and implements a sentence generation for an augmentive and alternative communication system(AAC). The AAC system is assistive communication device to help the mute language disorder communicate more freely and the system have an objected to reduce time and keystrokes for sentence generating. The paper of sentence generation make up for merits and demerits in the existing sentence generation method and in order to sentence generation. One aspect of Korean language that confines nouns defending on the verbs or postpositional words is used for sentence generation. The distinctive feature of this paper is to connect verbs to nouns using domain knowledge. We utilize the lexical information that exploits characteristics of Korean language for sentence generation. A comparison with other approaches is also presented. This sentence generation is based on lexical information by extracting characteristics of sentences.

  • PDF