• Title/Summary/Keyword: punctuation

Search Result 38, Processing Time 0.026 seconds

Analyzing the Defense Budgetary in the Republic of Korea with the Punctuated Equilibrium Theory (단절균형이론을 적용한 국방예산 분석에 관한 연구)

  • Yongjoon Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.6
    • /
    • pp.779-787
    • /
    • 2023
  • Previous research regarding budget analysis has been mostly limited to describing annual changes in defense budgets relative to total budgets without a theoretical background. More empirical defense budget research is needed with better data. This study conducts an empirical analysis of national defense expenditures using Punctuated Equilibrium Theory (PET). The purpose of this study is to examine trends in the Republic of Korea's (ROK) functional defense budgets (total defense budget, force operation budget, force improvement budget) and to identify and analyze radical points of change in the defense budget using punctuated equilibrium theory. This study also explores trends and punctuations in the national defense budgets using annual defense budget data from the ROK for every year from 1998 to 2017. This study finds that from 1998 to 2017 the spending pattern of the total defense budget in the ROK was characterized by 19 years of stable growth and a one-time punctuation (5.0%). The force operation budget exhibited stable growth in eighteen years and was punctuated twice (10%). The force improvement budget was punctuated five times.

Case Study on the Writing of the Papers of Journal of the Korean Association for Science Education (한국과학교육학회지 논문의 글쓰기 사례 연구)

  • Han, JaeYoung
    • Journal of The Korean Association For Science Education
    • /
    • v.35 no.4
    • /
    • pp.649-663
    • /
    • 2015
  • This study investigated the current state of writing in research papers of science education with focus on the translationese and basic Korean grammar, and found a way of improving the Korean language. The science education research have characteristics of both social science and natural science, and of having more quantitative than qualitative research, which could influence the writing of the research paper. The translationese means the conventional expression originated from foreign language other than Korean. The basic Korean grammar includes 'agreement,' 'spelling, word spacing, punctuation mark,' 'causative suffix,' 'use of English or loanword,' and the translationese is divided in 'English,' 'Japanese,' and 'English and Japanese.' The sentences in nine research papers in the 'Journal of the Korean Association for Science Education' were analyzed, and the problematic sentences were discussed and provided with alternatives. The cases with high frequency include '-jeok,' 'use of English,' 'expression of the plural,' 'passive voice of the verb with -hada,' '-go inneun,' '-eul tonghayeo,' '-e daehayeo,' 'gajida,' 'genitive case marker -eui,' 'passive voice with subject of thing,' and 'causative suffix, -sikida.' Based on the results, the characteristics of writing of science education research papers were described as 'writing of quantitative research,' 'objective writing of academic research,' and 'writing of research of foreign origin.' In order to improve the writing of research paper of science education, the science education researcher should pay attention to basic Korean grammar and the translationese, and be familiar with the concrete examples of problematic cases. The results of this study could be used in the education of writing and grammar of Korean language.

Analysis on Sentence Error Types of Mathematical Problem Posing of Pre-Service Elementary Teachers (초등학교 예비교사들의 수학적 '문제 만들기'에 나타나는 문장의 오류 유형 분석)

  • Huh, Nan;Shin, Hocheol
    • Journal of the Korean School Mathematics Society
    • /
    • v.16 no.4
    • /
    • pp.797-820
    • /
    • 2013
  • This study intended on analyzing the error patterns of mathematic problem posing sentences by the 100 elementary pre-teachers and discussing about the solutions. The results showed that the problem posing sentences have five error patterns: phonological error patterns, word error patterns, sentence error patterns, meaning error patterns, and notation error patterns. Divided into fourteen specific error patterns, they are as in the following. 1) Phonological error patterns are consisted of the 'ㄹ' addition error pattern and the abbreviated word error pattern. 2) Words error patterns are divided with the inappropriate usage of word error pattern and the inadequate abbreviation error pattern, which are formulized four subgroups such as the case maker, ending of the word, inappropriate usage of word, and inadequate abbreviation of article or word error pattern in detail. 3) Sentence error patterns are assumed four kinds of forms: the reference, ellipsis of sentence component, word order, and incomplete sentence error pattern. 4) Meaning error patterns are composed the logical contradiction and the ambiguous meaning. 5) Notation error patterns are formed four patterns as the spacing, punctuation, orthography of Hangul, and spelling rules of foreign words in Korean. Furthermore, the solutions for these error patterns were discussed: First, it has to be perceived the differences between spoken and written language. Second, it has to be rejected the spoken expressions in written contexts. Third, it should be focused on the learning of the basic sentence patterns during the class. Forth, it is suggested that the word meaning should have the logical development perception based on what it means. Finally, it is proposed that the system of spelling of Korean has to be learned. In addition to these suggestions, a new understanding is necessary regarding writing education for college students.

  • PDF

A Study on New material : (새 자료 <동방?이비겨리라> 연구)

  • Jo, Sang-Woo
    • (The)Study of the Eastern Classic
    • /
    • no.56
    • /
    • pp.75-115
    • /
    • 2014
  • The text reviewed in this paper "Dongbangsaek is the Secret (Dongbangsaegi bigyeorira)" is in the collection of the Yulgok Memorial Library of Dankook University. With 13 leaves ($35.7{\times}22.3cm$) bound with thread, the booklet has been transcribed by hand. Although there is no record on the place, person and year of transcription, it is estimated to have been transcribed in the 20th century based on the use of the period, a punctuation mark. In addition, the complete absence of dialect vocabulary also shows that it was transcribed in the capital area-Seoul or Gyeonggi Province. It is assumed that the text is part of a Buddhist scripture chanted by an exorcist during a shamanistic ritual. As a booklet containing secret methods to divine what is auspicious and what is ominous in daily life, it must have been transcribed by an exorcist to use it for her ritual.

Understanding Purposes and Functions of Students' Drawing while on Geological Field Trips and during Modeling-Based Learning Cycle (야외지질답사 및 모델링 기반 순환 학습에서 학생들이 그린 그림의 목적과 기능에 대한 이해)

  • Choi, Yoon-Sung
    • Journal of the Korean earth science society
    • /
    • v.42 no.1
    • /
    • pp.88-101
    • /
    • 2021
  • The purpose of this study was to qualitatively examine the meaning of students' drawings in outdoor classes and modeling-based learning cycles. Ten students were observed in a gifted education center in Seoul. Under the theme of the Hantan River, three outdoor classes and three modeling activities were conducted. Data were collected to document all student activities during field trips and classroom modeling activities using simultaneous video and audio recording and observation notes made by the researcher and students. Please note it is unclear what this citation refers to. If it is the previous sentence it should be placed within that sentence's punctuation. Hatisaru (2020) Ddrawing typess were classified by modifying the representations in a learning context in geological field trips. We used deductive content analysis to describe the drawing characteristics, including students writing. The results suggest that students have symbolic images that consist of geologic concepts, visual images that describe topographical features, and affective images that express students' emotion domains. The characteristics were classified into explanation, generality, elaboration, evidence, coherence, and state-of-mind. The characteristics and drawing types are consecutive in the modeling-based learning cycle and reflect the students' positive attitude and cognitive scientific domain. Drawing is a useful tool for reflecting students' thoughts and opinions in both outdoor class and classroom modeling activities. This study provides implications for emphasizing the importance of drawing activities.

Hate Speech Detection Using Modified Principal Component Analysis and Enhanced Convolution Neural Network on Twitter Dataset

  • Majed, Alowaidi
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.1
    • /
    • pp.112-119
    • /
    • 2023
  • Traditionally used for networking computers and communications, the Internet has been evolving from the beginning. Internet is the backbone for many things on the web including social media. The concept of social networking which started in the early 1990s has also been growing with the internet. Social Networking Sites (SNSs) sprung and stayed back to an important element of internet usage mainly due to the services or provisions they allow on the web. Twitter and Facebook have become the primary means by which most individuals keep in touch with others and carry on substantive conversations. These sites allow the posting of photos, videos and support audio and video storage on the sites which can be shared amongst users. Although an attractive option, these provisions have also culminated in issues for these sites like posting offensive material. Though not always, users of SNSs have their share in promoting hate by their words or speeches which is difficult to be curtailed after being uploaded in the media. Hence, this article outlines a process for extracting user reviews from the Twitter corpus in order to identify instances of hate speech. Through the use of MPCA (Modified Principal Component Analysis) and ECNN, we are able to identify instances of hate speech in the text (Enhanced Convolutional Neural Network). With the use of NLP, a fully autonomous system for assessing syntax and meaning can be established (NLP). There is a strong emphasis on pre-processing, feature extraction, and classification. Cleansing the text by removing extra spaces, punctuation, and stop words is what normalization is all about. In the process of extracting features, these features that have already been processed are used. During the feature extraction process, the MPCA algorithm is used. It takes a set of related features and pulls out the ones that tell us the most about the dataset we give itThe proposed categorization method is then put forth as a means of detecting instances of hate speech or abusive language. It is argued that ECNN is superior to other methods for identifying hateful content online. It can take in massive amounts of data and quickly return accurate results, especially for larger datasets. As a result, the proposed MPCA+ECNN algorithm improves not only the F-measure values, but also the accuracy, precision, and recall.

Financial Fraud Detection using Text Mining Analysis against Municipal Cybercriminality (지자체 사이버 공간 안전을 위한 금융사기 탐지 텍스트 마이닝 방법)

  • Choi, Sukjae;Lee, Jungwon;Kwon, Ohbyung
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.119-138
    • /
    • 2017
  • Recently, SNS has become an important channel for marketing as well as personal communication. However, cybercrime has also evolved with the development of information and communication technology, and illegal advertising is distributed to SNS in large quantity. As a result, personal information is lost and even monetary damages occur more frequently. In this study, we propose a method to analyze which sentences and documents, which have been sent to the SNS, are related to financial fraud. First of all, as a conceptual framework, we developed a matrix of conceptual characteristics of cybercriminality on SNS and emergency management. We also suggested emergency management process which consists of Pre-Cybercriminality (e.g. risk identification) and Post-Cybercriminality steps. Among those we focused on risk identification in this paper. The main process consists of data collection, preprocessing and analysis. First, we selected two words 'daechul(loan)' and 'sachae(private loan)' as seed words and collected data with this word from SNS such as twitter. The collected data are given to the two researchers to decide whether they are related to the cybercriminality, particularly financial fraud, or not. Then we selected some of them as keywords if the vocabularies are related to the nominals and symbols. With the selected keywords, we searched and collected data from web materials such as twitter, news, blog, and more than 820,000 articles collected. The collected articles were refined through preprocessing and made into learning data. The preprocessing process is divided into performing morphological analysis step, removing stop words step, and selecting valid part-of-speech step. In the morphological analysis step, a complex sentence is transformed into some morpheme units to enable mechanical analysis. In the removing stop words step, non-lexical elements such as numbers, punctuation marks, and double spaces are removed from the text. In the step of selecting valid part-of-speech, only two kinds of nouns and symbols are considered. Since nouns could refer to things, the intent of message is expressed better than the other part-of-speech. Moreover, the more illegal the text is, the more frequently symbols are used. The selected data is given 'legal' or 'illegal'. To make the selected data as learning data through the preprocessing process, it is necessary to classify whether each data is legitimate or not. The processed data is then converted into Corpus type and Document-Term Matrix. Finally, the two types of 'legal' and 'illegal' files were mixed and randomly divided into learning data set and test data set. In this study, we set the learning data as 70% and the test data as 30%. SVM was used as the discrimination algorithm. Since SVM requires gamma and cost values as the main parameters, we set gamma as 0.5 and cost as 10, based on the optimal value function. The cost is set higher than general cases. To show the feasibility of the idea proposed in this paper, we compared the proposed method with MLE (Maximum Likelihood Estimation), Term Frequency, and Collective Intelligence method. Overall accuracy and was used as the metric. As a result, the overall accuracy of the proposed method was 92.41% of illegal loan advertisement and 77.75% of illegal visit sales, which is apparently superior to that of the Term Frequency, MLE, etc. Hence, the result suggests that the proposed method is valid and usable practically. In this paper, we propose a framework for crisis management caused by abnormalities of unstructured data sources such as SNS. We hope this study will contribute to the academia by identifying what to consider when applying the SVM-like discrimination algorithm to text analysis. Moreover, the study will also contribute to the practitioners in the field of brand management and opinion mining.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.