• Title/Summary/Keyword: 문제해결 과정

Search Result 4,859, Processing Time 0.033 seconds

21st Century ROK's Art History Research on Central Eurasia (21세기 한국의 중앙유라시아 미술사 연구)

  • Lim, Young-ae
    • Korean Journal of Heritage: History & Science
    • /
    • v.48 no.3
    • /
    • pp.186-203
    • /
    • 2015
  • This article attempts to examine both the outcome and future task of the art history research on Central Eurasia, better known under the name of "Silk Road". The term Central Eurasia encompasses Xinjiang Uygur, Tibet, Mongolia, former Soviet Republics, the northwest region of India, Iran and Turkey. The article analyzes the 30-year history of the region's art history research and further presents a desirable direction that the study should move towards. Though short in its research period, the ROK's art history study on Central Eurasia has shown eye-catching achievement in several areas such as the northwest region of India and the Xinjiang Uygur, Dunhwang of China. Two factors allowed for this accomplishment. First was the actual improvement of the work environment, where the scholars were finally able to travel to Central Eurasia and explore the historic sites for themselves since 1990. More important was the 'arena of study' for the next-generation scholars made possible by institutions like The Korean Association for Central Asian Studies and the Center for Central Eurasian Studies. Slowly but consistently, the two academic societies induced scholars' attention towards the field and fostered new experts. Circumstances changed, marking 2012 as the starting point. International academic forums held by the government branches surged in number. The intention behind it was to link the ROK with the Silk Road and ultimately to obtain the "Eurasia initiative". As of now, the public has shown heightened interest in the issue. The academia is subsequently riding on this second "wave of interest" following the first wave in the 1980s. However, increased popularity comes with some negative consequences, and this art history research on Central Eurasia is no exception. There are criticisms regarding the objectivity of recent academic forums. Some argue that the aim of the forums are sternly set most of the times, prohibiting the presenters to voice their own perspectives. Still, this heated attention will definitely play its role as a stepping stone for further development. The academia should commit to fostering rising researchers who will systemically and professionally study the field. This is imperative in order for the Korean culture to successfully communicate with the world and take itself to a new level. Without completing this task, the ROK's art history research on Central Eurasia is likely to remain idle.

The Usage of the Vulgate Bible in the European Catholicism: from the Council of Trent until the Second Council of Vatican (유럽 천주교의 불가타 성경 사용 양상: 트렌토 공의회 이후부터 2차 바티칸 공의회 이전까지)

  • CHO, Hyeon Beom
    • The Critical Review of Religion and Culture
    • /
    • no.32
    • /
    • pp.257-287
    • /
    • 2017
  • It seems to be quite an ambitious endeavor to trace back the translation history of Catholic Vulgate Bible from Latin language to Asian languages since 16th century. I try to bring out the translation(translative) procedure of Latin Bible to the Chinese Version, which is eventually come up (and the latter)to the Korean Version. It has been supported and funded by the National Research Foundation of Korea. This task has a three-year plan. For the first step(operation), I examined and searched the European situation of the Vulgate Bible in the Catholic Church, i.e. the ritual use of Vulgate Bible in the Mass and the religious retreat. The liturgical texts, to begin with, were analysed to disclose how the Vulgate Bible was reflected in them. The Lectionary and the Evangeliary were the typical ones. The structure or the formation system of the Lectionaries for Mass was based on the liturgical year cycle. From this point, the Vulgate Bible was rooted in the religious life of European Catholics after the Council of Trent which had proclaimed the Vulgate to be authentic source of the Revelation, therefore, to be respected as the only authoritative Bible. How did the Catholic Church use the Vulgate Bible out of the context and the boundary (sphere) of liturgy? The Meditation guide books for the purpose of instructing the religious retreat was published and (diffused) circulated among the priests, the religious persons and even the laymen. In those books also were included (found) the citation, the interpretation and the commentaries of the Vulgate Bible. The most of the devotees in Europe read the biblical phrases out of the meditation guide books. There are still remained the unsolved problems of how to understand (for understanding) the actual aspect of the Vulgate Bible in the European Catholic Church. All the Biblical verses were translated into French and included in the meditation guide books published in France. What did the Holy See think the French translation of the Vulgate Bible? Unfortunately, there were not found the Vatican Decrees about the European translation of the Vulgate Bible. The relationship between the Vulgate Bible and the Meditation guide (Those) will be much important for the study of Chinese translation of it. The search for the Decrees and the researches on it and the European and the non-European translations of the Vulgate Bible will be a continuous task for me as well as the other researchers on these subjects in the future.

A Study on the Effect of the Document Summarization Technique on the Fake News Detection Model (문서 요약 기법이 가짜 뉴스 탐지 모형에 미치는 영향에 관한 연구)

  • Shim, Jae-Seung;Won, Ha-Ram;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.201-220
    • /
    • 2019
  • Fake news has emerged as a significant issue over the last few years, igniting discussions and research on how to solve this problem. In particular, studies on automated fact-checking and fake news detection using artificial intelligence and text analysis techniques have drawn attention. Fake news detection research entails a form of document classification; thus, document classification techniques have been widely used in this type of research. However, document summarization techniques have been inconspicuous in this field. At the same time, automatic news summarization services have become popular, and a recent study found that the use of news summarized through abstractive summarization has strengthened the predictive performance of fake news detection models. Therefore, the need to study the integration of document summarization technology in the domestic news data environment has become evident. In order to examine the effect of extractive summarization on the fake news detection model, we first summarized news articles through extractive summarization. Second, we created a summarized news-based detection model. Finally, we compared our model with the full-text-based detection model. The study found that BPN(Back Propagation Neural Network) and SVM(Support Vector Machine) did not exhibit a large difference in performance; however, for DT(Decision Tree), the full-text-based model demonstrated a somewhat better performance. In the case of LR(Logistic Regression), our model exhibited the superior performance. Nonetheless, the results did not show a statistically significant difference between our model and the full-text-based model. Therefore, when the summary is applied, at least the core information of the fake news is preserved, and the LR-based model can confirm the possibility of performance improvement. This study features an experimental application of extractive summarization in fake news detection research by employing various machine-learning algorithms. The study's limitations are, essentially, the relatively small amount of data and the lack of comparison between various summarization technologies. Therefore, an in-depth analysis that applies various analytical techniques to a larger data volume would be helpful in the future.

Esterification of Indonesia Tropical Crop Oil by Amberlyst-15 and Property Analysis of Biodiesel (인도네시아 열대작물 오일의 Amberlyst-15 촉매 에스테르화 반응 및 바이오디젤 물성 분석)

  • Lee, Kyoung-Ho;Lim, Riky;Lee, Joon-Pyo;Lee, Jin-Suk;Kim, Deog-Keun
    • Journal of the Korean Applied Science and Technology
    • /
    • v.36 no.1
    • /
    • pp.324-332
    • /
    • 2019
  • Most countries including Korea and Indonesia have strong policy for implementing biofuels like biodiesel. Shortage of the oil feedstock is the main barrier for increasing the supply of biodiesel fuel. In this study, in order to improve the stability of feedstock supply and lower the biodiesel production cost, the feasibility of biodiesel production using two types of Indonesian tropical crop oils, pressed at different harvesting times, were investigated. R. Trisperma oils, a high productive non-edible feedstocks, were investigated to produce biodiesel by esterification and transesterification because of it's high impurity and free fatty acid contents. the kindly provided oils from Indonesia were required to perform the filtering and water removal process to increase the efficiency of the esterificaton and transesterification reactions. The esterification used heterogeneous acid catalyst, Amberlyst-15. Before the reaction, the acid value of two types oil were 41, 17 mg KOH/g respectively. After the pre-esterification reaction, the acid value of oils were 3.7, 1.8 mg KOH/g respectively, the conversions were about 90%. Free fatty acid content was reduced to below 2%. Afterwards, the transesterification was performed using KOH as the base catalyst for transesterification. The prepared biodiesel showed about 93% of FAME content, and the total glycerol content was 0.43%. It did not meet the quality specification(FAME 96.5% and Total glycerol 0.24%) since the tested oils were identified to have a uncommon fatty acid, generally not found in vegetable oils, ${\alpha}$-eleostearic acid with much contents of 10.7~33.4%. So, it is required to perform the further research on reaction optimization and product purification to meet the fuel quality standards. So if the biodiesel production technology using un-utilized non-edible feedstock oils is successfully developed, stable supply of the feedstock for biodiesel production may be possible in the future.

Label Embedding for Improving Classification Accuracy UsingAutoEncoderwithSkip-Connections (다중 레이블 분류의 정확도 향상을 위한 스킵 연결 오토인코더 기반 레이블 임베딩 방법론)

  • Kim, Museong;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.175-197
    • /
    • 2021
  • Recently, with the development of deep learning technology, research on unstructured data analysis is being actively conducted, and it is showing remarkable results in various fields such as classification, summary, and generation. Among various text analysis fields, text classification is the most widely used technology in academia and industry. Text classification includes binary class classification with one label among two classes, multi-class classification with one label among several classes, and multi-label classification with multiple labels among several classes. In particular, multi-label classification requires a different training method from binary class classification and multi-class classification because of the characteristic of having multiple labels. In addition, since the number of labels to be predicted increases as the number of labels and classes increases, there is a limitation in that performance improvement is difficult due to an increase in prediction difficulty. To overcome these limitations, (i) compressing the initially given high-dimensional label space into a low-dimensional latent label space, (ii) after performing training to predict the compressed label, (iii) restoring the predicted label to the high-dimensional original label space, research on label embedding is being actively conducted. Typical label embedding techniques include Principal Label Space Transformation (PLST), Multi-Label Classification via Boolean Matrix Decomposition (MLC-BMaD), and Bayesian Multi-Label Compressed Sensing (BML-CS). However, since these techniques consider only the linear relationship between labels or compress the labels by random transformation, it is difficult to understand the non-linear relationship between labels, so there is a limitation in that it is not possible to create a latent label space sufficiently containing the information of the original label. Recently, there have been increasing attempts to improve performance by applying deep learning technology to label embedding. Label embedding using an autoencoder, a deep learning model that is effective for data compression and restoration, is representative. However, the traditional autoencoder-based label embedding has a limitation in that a large amount of information loss occurs when compressing a high-dimensional label space having a myriad of classes into a low-dimensional latent label space. This can be found in the gradient loss problem that occurs in the backpropagation process of learning. To solve this problem, skip connection was devised, and by adding the input of the layer to the output to prevent gradient loss during backpropagation, efficient learning is possible even when the layer is deep. Skip connection is mainly used for image feature extraction in convolutional neural networks, but studies using skip connection in autoencoder or label embedding process are still lacking. Therefore, in this study, we propose an autoencoder-based label embedding methodology in which skip connections are added to each of the encoder and decoder to form a low-dimensional latent label space that reflects the information of the high-dimensional label space well. In addition, the proposed methodology was applied to actual paper keywords to derive the high-dimensional keyword label space and the low-dimensional latent label space. Using this, we conducted an experiment to predict the compressed keyword vector existing in the latent label space from the paper abstract and to evaluate the multi-label classification by restoring the predicted keyword vector back to the original label space. As a result, the accuracy, precision, recall, and F1 score used as performance indicators showed far superior performance in multi-label classification based on the proposed methodology compared to traditional multi-label classification methods. This can be seen that the low-dimensional latent label space derived through the proposed methodology well reflected the information of the high-dimensional label space, which ultimately led to the improvement of the performance of the multi-label classification itself. In addition, the utility of the proposed methodology was identified by comparing the performance of the proposed methodology according to the domain characteristics and the number of dimensions of the latent label space.

A Study of the Application of 'Digital Heritage ODA' - Focusing on the Myanmar cultural heritage management system - (디지털 문화유산 ODA 적용에 관한 시론적 연구 -미얀마 문화유산 관리시스템을 중심으로-)

  • Jeong, Seongmi
    • Korean Journal of Heritage: History & Science
    • /
    • v.53 no.4
    • /
    • pp.198-215
    • /
    • 2020
  • Official development assistance refers to assistance provided by governments and other public institutions in donor countries, aimed at promoting economic development and social welfare in developing countries. The purpose of this research is to examine the construction process of the "Myanmar Cultural Heritage Management System" that is underway as part of the ODA project to strengthen cultural and artistic capabilities and analyze the achievements and challenges of the Digital Cultural Heritage ODA. The digital cultural heritage management system is intended to achieve the permanent preservation and sustainable utilization of tangible and intangible cultural heritage materials. Cultural heritage can be stored in digital archives, newly approached using computer analysis technology, and information can be used in multiple dimensions. First, the Digital Cultural Heritage ODA was able to permanently preserve cultural heritage content that urgently needed digitalization by overcoming and documenting the "risk" associated with cultural heritage under threat of being extinguished, damaged, degraded, or distorted in Myanmar. Second, information on Myanmar's cultural heritage can be systematically managed and used in many ways through linkages between materials. Third, cultural maps can be implemented that are based on accurate geographical location information as to where cultural heritage is located or inherited. Various items of cultural heritage were collectively and intensively visualized to maximize utility and convenience for academic, policy, and practical purposes. Fourth, we were able to overcome the one-sided limitations of cultural ODA in relations between donor and recipient countries. Fifth, the capacity building program run by officials in charge of the beneficiary country, which could be the most important form of sustainable development in the cultural ODA, was operated together. Sixth, there is an implication that it is an ODA that can be relatively smooth and non-face-to-face in nature, without requiring the movement of manpower between countries during the current global pandemic. However, the following tasks remain to be solved through active discussion and deliberation in the future. First, the content of the data uploaded to the system should be verified. Second, to preserve digital cultural heritage, it must be protected from various threats. For example, it is necessary to train local experts to prepare for errors caused by computer viruses, stored data, or operating systems. Third, due to the nature of the rapidly changing environment of computer technology, measures should also be discussed to address the problems that tend to follow when new versions and programs are developed after the end of the ODA project, or when developers have not continued to manage their programs. Fourth, since the classification system criteria and decisions regarding whether the data will be disclosed or not are set according to Myanmar's political judgment, it is necessary to let the beneficiary country understand the ultimate purpose of the cultural ODA project.

The Applicability of Conditional Generative Model Generating Groundwater Level Fluctuation Corresponding to Precipitation Pattern (조건부 생성모델을 이용한 강수 패턴에 따른 지하수위 생성 및 이의 활용에 관한 연구)

  • Jeong, Jiho;Jeong, Jina;Lee, Byung Sun;Song, Sung-Ho
    • Economic and Environmental Geology
    • /
    • v.54 no.1
    • /
    • pp.77-89
    • /
    • 2021
  • In this study, a method has been proposed to improve the performance of hydraulic property estimation model developed by Jeong et al. (2020). In their study, low-dimensional features of the annual groundwater level (GWL) fluctuation patterns extracted based on a Denoising autoencoder (DAE) was used to develop a regression model for predicting hydraulic properties of an aquifer. However, low-dimensional features of the DAE are highly dependent on the precipitation pattern even if the GWL is monitored at the same location, causing uncertainty in hydraulic property estimation of the regression model. To solve the above problem, a process for generating the GWL fluctuation pattern for conditioning the precipitation is proposed based on a conditional variational autoencoder (CVAE). The CVAE trains a statistical relationship between GWL fluctuation and precipitation pattern. The actual GWL and precipitation data monitored on a total of 71 monitoring stations over 10 years in South Korea was applied to validate the effect of using CVAE. As a result, the trained CVAE model reasonably generated GWL fluctuation pattern with the conditioning of various precipitation patterns for all the monitoring locations. Based on the trained CVAE model, the low-dimensional features of the GWL fluctuation pattern without interference of different precipitation patterns were extracted for all monitoring stations, and they were compared to the features extracted based on the DAE. Consequently, it can be confirmed that the statistical consistency of the features extracted using CVAE is improved compared to DAE. Thus, we conclude that the proposed method may be useful in extracting a more accurate feature of GWL fluctuation pattern affected solely by hydraulic characteristics of the aquifer, which would be followed by the improved performance of the previously developed regression model.

The Myth of Huang-ti(the Yellow Emperor) and the Construction of Chinese Nationhood in Late Qing(淸) ("나의 피 헌원(軒轅)에 바치리라" - 황제신화(黃帝神話)와 청말(淸末) '네이션(민족)' 구조의 확립 -)

  • Shen, Sung-chaio;Jo, U-Yeon
    • Journal of Korean Historical Folklife
    • /
    • no.27
    • /
    • pp.267-361
    • /
    • 2008
  • This article traces how the modern Chinese "nation" was constructed as an "imagined community" around Huang-ti (the Yellow Emperor) in late Qing. Huang-ti was a legendary figure in ancient China and the imperial courts monopolized the worship of him. Many late Qing intellectuals appropriated this symbolic figure and, through a set of discursive strategies of "framing, voice and narrative structure," transformed him into a privileged symbol for modern Chinese national identity. What Huang-ti could offer was, however, no more than a "public face" for the imagined new national community, or in other words, a formal structure without substantial contents. No consensus appeared on whom the Chinese nation should include and where the Chinese nation should draw its boundaries. The anti-Manchu revolutionaries emphasized the primordial attachment of blood and considered modern China an exclusive community of Huang-ti's descent. The constitutional reformers sought to stretch the boundaries to include the ethnic groups other than the Han. Some minority intellectuals, particularly the Manchu ones, re-constructed the historic memory of their ethnic origin around Huang-ti. The quarrels among intellectuals of different political persuasion testify how Huang-ti as the most powerful cultural symbol became a site for contests and negotiations in the late Qing process of national construction.

A Study on the Importance and Priorities of the Investment Determinants of Startup Accelerators (스타트업 액셀러레이터 투자결정요인의 중요도 및 우선순위에 대한 연구)

  • Heo, Joo-yeun
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.15 no.6
    • /
    • pp.27-42
    • /
    • 2020
  • Startup accelerators have emerged as new investment entities that help early startups, which are not easy to survive continuously due to lack of funds, commercialization capabilities, and experiences. As their positive performance on early startups and the ecosystem has been proven, the number of early startups which want to receive their investment is also increasing. However, they are vaguely preparing to attract accelerators' investment because they do not have any information on what factors the accelerators consider important. In addition, researches on startup accelerators are also at an early level, so there are no remarkable prior studies on factors that decide on investment. Therefore, this study aims to help startups prepare for investment attraction by looking at what factors are important for accelerators to invest, and to provide meaningful implications to academia. In the preceding study, we derived five upper level categories, 26 lower level accelerators' investment determinants through the qualitative meta-synthesis method, secondary data analysis, observation on US accelerators and in-depth interviews. In this study, we want to derive important implications by deriving priorities of the accelerators' investment determinants. Therefore, we used AHP that are evaluated as the suitable methodology for deriving importance and priority. The analysis results show that accelerators value market-related factors most. This means that startups that are subject to investment by accelerators are early-stage startups, and many companies have not fully developed their products or services. Therefore, market-related factors that can be evaluated objectively seem to be more important than products (or services) that are still ambiguous. Next, it was found that the factors related to the internal workforce of startups are more important. Since accelerators want to develop their businesses together with start-ups and team members through mentoring, ease of collaboration with them is very important, which seems to be important. The overall priority analysis results of the 26 investment determinants show that 'customer needs' and 'founders and team members' understanding of customers and markets' (0.62) are important and high priority factors. The results also show that startup accelerators consider the customer-centered perspective very important. And among the factors related to startups, the most prominent factor was the founder's openness and execution ability. Therefore, it can be confirmed that accelerators consider the ease of collaboration with these startups very important.

KB-BERT: Training and Application of Korean Pre-trained Language Model in Financial Domain (KB-BERT: 금융 특화 한국어 사전학습 언어모델과 그 응용)

  • Kim, Donggyu;Lee, Dongwook;Park, Jangwon;Oh, Sungwoo;Kwon, Sungjun;Lee, Inyong;Choi, Dongwon
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.2
    • /
    • pp.191-206
    • /
    • 2022
  • Recently, it is a de-facto approach to utilize a pre-trained language model(PLM) to achieve the state-of-the-art performance for various natural language tasks(called downstream tasks) such as sentiment analysis and question answering. However, similar to any other machine learning method, PLM tends to depend on the data distribution seen during the training phase and shows worse performance on the unseen (Out-of-Distribution) domain. Due to the aforementioned reason, there have been many efforts to develop domain-specified PLM for various fields such as medical and legal industries. In this paper, we discuss the training of a finance domain-specified PLM for the Korean language and its applications. Our finance domain-specified PLM, KB-BERT, is trained on a carefully curated financial corpus that includes domain-specific documents such as financial reports. We provide extensive performance evaluation results on three natural language tasks, topic classification, sentiment analysis, and question answering. Compared to the state-of-the-art Korean PLM models such as KoELECTRA and KLUE-RoBERTa, KB-BERT shows comparable performance on general datasets based on common corpora like Wikipedia and news articles. Moreover, KB-BERT outperforms compared models on finance domain datasets that require finance-specific knowledge to solve given problems.