• Title/Summary/Keyword: 언어TEXT

Search Result 757, Processing Time 0.04 seconds

Multi-Emotion Regression Model for Recognizing Inherent Emotions in Speech Data (음성 데이터의 내재된 감정인식을 위한 다중 감정 회귀 모델)

  • Moung Ho Yi;Myung Jin Lim;Ju Hyun Shin
    • Smart Media Journal
    • /
    • v.12 no.9
    • /
    • pp.81-88
    • /
    • 2023
  • Recently, communication through online is increasing due to the spread of non-face-to-face services due to COVID-19. In non-face-to-face situations, the other person's opinions and emotions are recognized through modalities such as text, speech, and images. Currently, research on multimodal emotion recognition that combines various modalities is actively underway. Among them, emotion recognition using speech data is attracting attention as a means of understanding emotions through sound and language information, but most of the time, emotions are recognized using a single speech feature value. However, because a variety of emotions exist in a complex manner in a conversation, a method for recognizing multiple emotions is needed. Therefore, in this paper, we propose a multi-emotion regression model that extracts feature vectors after preprocessing speech data to recognize complex, inherent emotions and takes into account the passage of time.

Implementation of a Scheme Mobile Programming Application and Performance Evaluation of the Interpreter (Scheme 프로그래밍 모바일 앱 구현과 인터프리터 성능 평가)

  • Dongseob Kim;Sangkon Han;Gyun Woo
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.3
    • /
    • pp.122-129
    • /
    • 2024
  • Though programming education has been stressed recently, the elementary, middle, and high school students are having trouble in programming education. Most programming environments for them are based on block coding, which hinders them from moving to text coding. The traditional PC environment has also troubles such as maintenance problems. In this situation, mobile applications can be considered as alternative programming environments. This paper addresses the design and implementation of coding applications for mobile devices. As a prototype, a Scheme interpreter mobile app is proposed, where Scheme is used for programming courses at MIT since it supports multi-paradigm programming. The implementation has the advantage of not consuming the network bandwidth since it is designed as a standalone application. According to the benchmark result, the execution time on Android devices, relative to that on a desktop, was 131% for the Derivative and 157% for the Tak. Further, the maximum execution times for the benchmark programs on the Android device were 19.8ms for the Derivative and 131.15ms for the Tak benchmark. This confirms that when selecting an Android device for programming education purposes, there are no significant constraints for training.

Research on ITB Contract Terms Classification Model for Risk Management in EPC Projects: Deep Learning-Based PLM Ensemble Techniques (EPC 프로젝트의 위험 관리를 위한 ITB 문서 조항 분류 모델 연구: 딥러닝 기반 PLM 앙상블 기법 활용)

  • Hyunsang Lee;Wonseok Lee;Bogeun Jo;Heejun Lee;Sangjin Oh;Sangwoo You;Maru Nam;Hyunsik Lee
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.11
    • /
    • pp.471-480
    • /
    • 2023
  • The Korean construction order volume in South Korea grew significantly from 91.3 trillion won in public orders in 2013 to a total of 212 trillion won in 2021, particularly in the private sector. As the size of the domestic and overseas markets grew, the scale and complexity of EPC (Engineering, Procurement, Construction) projects increased, and risk management of project management and ITB (Invitation to Bid) documents became a critical issue. The time granted to actual construction companies in the bidding process following the EPC project award is not only limited, but also extremely challenging to review all the risk terms in the ITB document due to manpower and cost issues. Previous research attempted to categorize the risk terms in EPC contract documents and detect them based on AI, but there were limitations to practical use due to problems related to data, such as the limit of labeled data utilization and class imbalance. Therefore, this study aims to develop an AI model that can categorize the contract terms based on the FIDIC Yellow 2017(Federation Internationale Des Ingenieurs-Conseils Contract terms) standard in detail, rather than defining and classifying risk terms like previous research. A multi-text classification function is necessary because the contract terms that need to be reviewed in detail may vary depending on the scale and type of the project. To enhance the performance of the multi-text classification model, we developed the ELECTRA PLM (Pre-trained Language Model) capable of efficiently learning the context of text data from the pre-training stage, and conducted a four-step experiment to validate the performance of the model. As a result, the ensemble version of the self-developed ITB-ELECTRA model and Legal-BERT achieved the best performance with a weighted average F1-Score of 76% in the classification of 57 contract terms.

A Study on the Rhythm of Sijo Using Prosodie Analysis - Centering on < Ouga > by Seon-do Yun - (프로조디(prosodie) 분석을 통한 시조의 가락 고찰 시론(試論) - 윤선도(尹善道)의 <오우가(五友歌)>를 대상으로 -)

  • Kim, Seong-Moon
    • Sijohaknonchong
    • /
    • v.43
    • /
    • pp.41-66
    • /
    • 2015
  • A study on rhythm of a sijo was mostly conducted based on rhythm theory. As it is considered to define the rhythm of a formal sijo based on three verses, its significance has been recognized. However, if rhythm is understood to be superior to cadence or versification, it seems necessary to examine the rhythm of a sijo as a verse with a fixed form as well as a highly individual rhythm of each and every lyric poet, which is informal rhythm, in order to fully understand them. In this case, prosodie analysis by H. Meschonnic (1932~ 2009) can be a significant methodology. As this study gropes for a possibility to examine the rhythm of a sijo from a new perspective instead of existing rhythm theory through the application of H. Meschonnic's prosodie analysis, it can be regarded as an essay. Prosodie newly suggested by Meschonnic is referred to as linguistic organization of consonants and vowels and indication of their paradigm, and it conflicts the perspective that traditionally separates linguistic sound from meaning for dichotomous understanding. It is due to the fact that the organization of consonants and vowels is a unit that constitutes a complicated layer of significant sound and meaning. Accordingly, prosodie analysis that is irregularly and aperiodically distributed within poetic text can be considered as methodology aimed at explaining how a poem is integrated in terms of sound and semantics. The core of prosodie analysis is to examine how the phonologic system stands against the theme of a poem. It ultimately has the same way of establishing literary style of a poet as it is to explain a unique aesthetic structure that individual poems have and show distinct characteristics of linguistic use by a poet. Prior to application of the prosodie analysis to sijo in general, the study preparatorily conducted prosodie analysis on < Ouga > by Gosan Seon-do Yun.

  • PDF

Deep Learning OCR based document processing platform and its application in financial domain (금융 특화 딥러닝 광학문자인식 기반 문서 처리 플랫폼 구축 및 금융권 내 활용)

  • Dongyoung Kim;Doohyung Kim;Myungsung Kwak;Hyunsoo Son;Dongwon Sohn;Mingi Lim;Yeji Shin;Hyeonjung Lee;Chandong Park;Mihyang Kim;Dongwon Choi
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.1
    • /
    • pp.143-174
    • /
    • 2023
  • With the development of deep learning technologies, Artificial Intelligence powered Optical Character Recognition (AI-OCR) has evolved to read multiple languages from various forms of images accurately. For the financial industry, where a large number of diverse documents are processed through manpower, the potential for using AI-OCR is great. In this study, we present a configuration and a design of an AI-OCR modality for use in the financial industry and discuss the platform construction with application cases. Since the use of financial domain data is prohibited under the Personal Information Protection Act, we developed a deep learning-based data generation approach and used it to train the AI-OCR models. The AI-OCR models are trained for image preprocessing, text recognition, and language processing and are configured as a microservice architected platform to process a broad variety of documents. We have demonstrated the AI-OCR platform by applying it to financial domain tasks of document sorting, document verification, and typing assistance The demonstrations confirm the increasing work efficiency and conveniences.

Sentiment Analysis of Korean Reviews Using CNN: Focusing on Morpheme Embedding (CNN을 적용한 한국어 상품평 감성분석: 형태소 임베딩을 중심으로)

  • Park, Hyun-jung;Song, Min-chae;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.59-83
    • /
    • 2018
  • With the increasing importance of sentiment analysis to grasp the needs of customers and the public, various types of deep learning models have been actively applied to English texts. In the sentiment analysis of English texts by deep learning, natural language sentences included in training and test datasets are usually converted into sequences of word vectors before being entered into the deep learning models. In this case, word vectors generally refer to vector representations of words obtained through splitting a sentence by space characters. There are several ways to derive word vectors, one of which is Word2Vec used for producing the 300 dimensional Google word vectors from about 100 billion words of Google News data. They have been widely used in the studies of sentiment analysis of reviews from various fields such as restaurants, movies, laptops, cameras, etc. Unlike English, morpheme plays an essential role in sentiment analysis and sentence structure analysis in Korean, which is a typical agglutinative language with developed postpositions and endings. A morpheme can be defined as the smallest meaningful unit of a language, and a word consists of one or more morphemes. For example, for a word '예쁘고', the morphemes are '예쁘(= adjective)' and '고(=connective ending)'. Reflecting the significance of Korean morphemes, it seems reasonable to adopt the morphemes as a basic unit in Korean sentiment analysis. Therefore, in this study, we use 'morpheme vector' as an input to a deep learning model rather than 'word vector' which is mainly used in English text. The morpheme vector refers to a vector representation for the morpheme and can be derived by applying an existent word vector derivation mechanism to the sentences divided into constituent morphemes. By the way, here come some questions as follows. What is the desirable range of POS(Part-Of-Speech) tags when deriving morpheme vectors for improving the classification accuracy of a deep learning model? Is it proper to apply a typical word vector model which primarily relies on the form of words to Korean with a high homonym ratio? Will the text preprocessing such as correcting spelling or spacing errors affect the classification accuracy, especially when drawing morpheme vectors from Korean product reviews with a lot of grammatical mistakes and variations? We seek to find empirical answers to these fundamental issues, which may be encountered first when applying various deep learning models to Korean texts. As a starting point, we summarized these issues as three central research questions as follows. First, which is better effective, to use morpheme vectors from grammatically correct texts of other domain than the analysis target, or to use morpheme vectors from considerably ungrammatical texts of the same domain, as the initial input of a deep learning model? Second, what is an appropriate morpheme vector derivation method for Korean regarding the range of POS tags, homonym, text preprocessing, minimum frequency? Third, can we get a satisfactory level of classification accuracy when applying deep learning to Korean sentiment analysis? As an approach to these research questions, we generate various types of morpheme vectors reflecting the research questions and then compare the classification accuracy through a non-static CNN(Convolutional Neural Network) model taking in the morpheme vectors. As for training and test datasets, Naver Shopping's 17,260 cosmetics product reviews are used. To derive morpheme vectors, we use data from the same domain as the target one and data from other domain; Naver shopping's about 2 million cosmetics product reviews and 520,000 Naver News data arguably corresponding to Google's News data. The six primary sets of morpheme vectors constructed in this study differ in terms of the following three criteria. First, they come from two types of data source; Naver news of high grammatical correctness and Naver shopping's cosmetics product reviews of low grammatical correctness. Second, they are distinguished in the degree of data preprocessing, namely, only splitting sentences or up to additional spelling and spacing corrections after sentence separation. Third, they vary concerning the form of input fed into a word vector model; whether the morphemes themselves are entered into a word vector model or with their POS tags attached. The morpheme vectors further vary depending on the consideration range of POS tags, the minimum frequency of morphemes included, and the random initialization range. All morpheme vectors are derived through CBOW(Continuous Bag-Of-Words) model with the context window 5 and the vector dimension 300. It seems that utilizing the same domain text even with a lower degree of grammatical correctness, performing spelling and spacing corrections as well as sentence splitting, and incorporating morphemes of any POS tags including incomprehensible category lead to the better classification accuracy. The POS tag attachment, which is devised for the high proportion of homonyms in Korean, and the minimum frequency standard for the morpheme to be included seem not to have any definite influence on the classification accuracy.

A Design and Implementation of Event Processor for Playing SMIL 2.0 Documents (SMIL 2.0 문서 재생을 위한 이벤트 처리기의 설계 및 구현)

  • 김혜은;채진석;이재원;김성동;이종우
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.2
    • /
    • pp.251-263
    • /
    • 2004
  • The Synchronized Multimedia Integration Language (SMIL), recommended by the World Wide Web Consortium (W3C) in 1998, is an XML-based declarative language to synchronize and present multimedia documents. SMIL can create new multimedia data integrating various types of multimedia objects which exist separately such as text, video, graphics and audio. It can support synchronization of multimedia data which are limited in current HTML-based Web technology. For its popularity, it is required to develop a multimedia server guaranteeing Quality of Service (QoS), authoring tool and player. For developing a SMIL authoring tool and player, the technologies are essentially required to read and analyze a SMIL document and to play synchronized various types of media objects in a timeline. In this paper, we describe a design and implementation of an event processor which supports SMIL 2.0 timing model. Moreover, we also develop a SMIL 2.0 player using the proposed event processor. This will facilitate the play of SMIL contents, so that it can contribute to the prosperity of SMIL technology It is possible to reuse in various language profiles defined in the SMIL standard. This player is expected to be utilized in other standard integrating SMIL such as XHTML+SMIL and SMIL Animation.

  • PDF

Development of a gridded crop growth simulation system for the DSSAT model using script languages (스크립트 언어를 사용한 DSSAT 모델 기반 격자형 작물 생육 모의 시스템 개발)

  • Yoo, Byoung Hyun;Kim, Kwang Soo;Ban, Ho-Young
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.20 no.3
    • /
    • pp.243-251
    • /
    • 2018
  • The gridded simulation of crop growth, which would be useful for shareholders and policy makers, often requires specialized computation tasks for preparation of weather input data and operation of a given crop model. Here we developed an automated system to allow for crop growth simulation over a region using the DSSAT (Decision Support System for Agrotechnology Transfer) model. The system consists of modules implemented using R and shell script languages. One of the modules has a functionality to create weather input files in a plain text format for each cell. Another module written in R script was developed for GIS data processing and parallel computing. The other module that launches the crop model automatically was implemented using the shell script language. As a case study, the automated system was used to determine the maximum soybean yield for a given set of management options in Illinois state in the US. The AgMERRA dataset, which is reanalysis data for agricultural models, was used to prepare weather input files during 1981 - 2005. It took 7.38 hours to create 1,859 weather input files for one year of soybean growth simulation in Illinois using a single CPU core. In contrast, the processing time decreased considerably, e.g., 35 minutes, when 16 CPU cores were used. The automated system created a map of the maturity group and the planting date that resulted in the maximum yield in a raster data format. Our results indicated that the automated system for the DSSAT model would help spatial assessments of crop yield at a regional scale.

Exploratory Study on the Possibilities of Convergence with Music in Writing Classes (글쓰기 수업에서 음악과의 융합 가능성에 대한 탐색적 연구)

  • Lee, Ran
    • The Journal of the Korea Contents Association
    • /
    • v.20 no.8
    • /
    • pp.88-100
    • /
    • 2020
  • This is an exploratory study based on the literature reviews which examined the possibilities and necessities of multimodal writing curriculum for liberal education. The purpose of this study is to analyze the existing research results which utilized the teaching methods associating music and writing, and to find the educational implications, and finally in terms of writing education, to suggest the possibilities of writing classes' convergent forms with music extracted from the results of the existing studies. Those studies were categorized to four patterns: WAC, effects of therapy, materials for writing, and new literacy. Based on Meyrowitz's perspective, firstly music can be utilized as a circumstance, which means a teacher can indirectly take the emotional, reminding, and healing effects of background musics. Secondly, music can play an important role of materials in thinking and writing, which is the most generally utilized pattern today. The effects are found in all of affective, cognitive, and strategic domains by utilizing music as a sort of reading materials. Thirdly, the convergent writing of music and narrative is suggested. Music is an independent language that can interact with narrative and construct text meanings in this kind of writing classes. These three dimensions of convergence have different perspectives, but sometimes occur at a same time or as a connected pattern. This study proposes that writing teachers need to improve their competence in music as well and to have professional concerns and efforts to develop their convergent writing teaching skills with music for these classes. Finally, this study stresses that team teaching can be an alternative for them.

Multi-stage News Classification System for Predicting Stock Price Changes (주식 가격 변동 예측을 위한 다단계 뉴스 분류시스템)

  • Paik, Woo-Jin;Kyung, Myoung-Hyoun;Min, Kyung-Soo;Oh, Hye-Ran;Lim, Cha-Mi;Shin, Moon-Sun
    • Journal of the Korean Society for information Management
    • /
    • v.24 no.2
    • /
    • pp.123-141
    • /
    • 2007
  • It has been known that predicting stock price is very difficult due to a large number of known and unknown factors and their interactions, which could influence the stock price. However, we started with a simple assumption that good news about a particular company will likely to influence its stock price to go up and vice versa. This assumption was verified to be correct by manually analyzing how the stock prices change after the relevant news stories were released. This means that we will be able to predict the stock price change to a certain degree if there is a reliable method to classify news stories as either favorable or unfavorable toward the company mentioned in the news. To classify a large number of news stories consistently and rapidly, we developed and evaluated a natural language processing based multi-stage news classification system, which categorizes news stories into either good or bad. The evaluation result was promising as the automatic classification led to better than chance prediction of the stock price change.