• Title/Summary/Keyword: 백과사전적 지식

Search Result 24, Processing Time 0.029 seconds

Knowledge Structure and Terminology of Archival Appraisal: Focusing on the Entries of "Encyclopedia of Archival Science" (기록평가의 지식구조와 용어 분석 : 『Encyclopedia of Archival Science』 표제어를 중심으로)

  • Moon-Won Seol
    • Journal of Korean Society of Archives and Records Management
    • /
    • v.23 no.1
    • /
    • pp.81-99
    • /
    • 2023
  • Archival appraisal is a key area of archival science. This study aims to identify the knowledge structure of archival appraisal and propose appropriate translation terms. To this end, 34 entry terms related to archival appraisal were first extracted from 146 entries in the Encyclopedia of Archival Science, a leading encyclopedia in the field, and terminologies were categorized based on content analysis. Second, the contents described in the entries were assessed for each category, and the main concepts and flow of discussions were also examined. Third and last, suitable translations for each entry term were presented, and considerations were suggested for the translations, such as the application of appropriate principles of equivalence. Specifically, terms such as "records appraisal" and "archival appraisal" were recommended to be used separately.

Extracting English-Korean Named-Entity Word-pairs using Wikipedia (위키피디아를 이용한 영-한 개체명 대역어 쌍 구축)

  • Kim, Eun-Kyung;Choi, Key-Sun
    • Annual Conference on Human and Language Technology
    • /
    • 2009.10a
    • /
    • pp.101-105
    • /
    • 2009
  • 본 논문은 공통적으로 이용할 수 있는 웹 환경에서의 한국어 정보로 획득할 수 있는 정보의 양이 영어권 정보의 양보다 상대적으로 적다는 것을 토대로, 웹정보 이용의 불균형을 해소하고자 하는 목적으로부터 출발하였다. 최근에는 지식 정보의 세계화, 국제화에 따라 동일한 정보를 각국 언어로 제공하고자하는 연구가 꾸준히 증가하고 있다. 온라인 백과사전인 위키피디아 역시 현재 다국어로 제공이 되고 있지만 한국어로 작성된 문서는 영어로 작성된 문서의 5% 미만인 것으로 조사되었다. 본 논문에서는 위키피디아 내에서 제공하는 다국어간의 링크 정보와 인포박스 데이터를 활용하여 위키피디아 문서 내에서 개체명을 인식하고, 자동으로 개체명의 영-한 대역어 쌍을 추출하는 것을 목표로 한다. 개체명은 일반 사전에 등재 되지 않은 경우가 많기 때문에, 기계번역에서 사전 데이터 등을 활용하여 개체명을 처리하는 것은 쉽지 않으며 일반적으로 음차표기 방식을 함께 사용하여 해결하고 있다. 본 논문을 통해 위키피디아 데이터를 활용해 만들어진 영-한 개체명 대역어 사전을 구축하기 위해 사용된 기술은 추후 위키피디아 문서를 기계번역하는데 있어 동일한 방법으로 사용이 가능하며, 구축된 사전 데이터는 추후 영-한 자동 음차표기 연구의 사전 데이터로도 활용이 가능하다.

  • PDF

Natural Language Processing and Cognition (자연언어처리와 인지)

  • 이정민
    • Korean Journal of Cognitive Science
    • /
    • v.3 no.2
    • /
    • pp.161-174
    • /
    • 1992
  • The present discussion is concerned with showing the development of natural language processing and how it is related to information and cognition.On the basis of the computeational model,in which humans are viewed as processors of linguistic structures that use stored knowledge-grammar, lexicon and structures representing the encyclopedic information of the world,such programs of natural language understanding as Winograd's SHRDLU came out.However,such pragmatic factors as contexts and the speaker's beliefs,internts,goals and intentions are not easy to process yet.Language,ingormation and cognition are argued to be closely interrelated,and the study of them,the paper argues,can lead to the development of science on general.

Politics of Collective Intelligence - Paradigm Shift of Knowledge and its Possibility on Democracy - (집단지성의 정치 - 지식패러다임의 변화와 민주주의의 가능성 -)

  • Jho, Whasun;Cho, Jaedong
    • Informatization Policy
    • /
    • v.17 no.4
    • /
    • pp.61-79
    • /
    • 2010
  • This study focuses on the emergence of collective intelligence and its impact on the democracy in the information era. Scholars have posed very different-optimistic and pessimistic-views on the possibility of collective knowledge produced by the public. Focusing on the cases of a free online encylopedia known as wikipedia and 2008 Candlelight Demonstration against the imports of US beef in Korea, this paper analyzes the mechanism of collective intelligence and its political implications on the democracy. Specifically, this article approaches changes in new knowledge paradigm with two different variables: the degree of connectivity and the quality of deliberation. Applying two different sets of variables helps us to distinguish the possibilities of collective intelligence and anti-intelligence, which would suggest social and political implications for the democracy in a country. This study finds a critical difference in terms of the quality of deliberation, measured by the indicators such as diversity, independence, and integration mechanism for online deliberation.

  • PDF

Design and Implementation of a Hypermedia System for Effective Multimedia Information Retrieval (멀티미디어 정보의 효율적인 검색을 위한 하이퍼미디어 시스템의 설계와 구현)

  • 고영곤;최윤철
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.18 no.8
    • /
    • pp.1213-1225
    • /
    • 1993
  • Hypermedia systems have the browsing mechanism using links and provide navigation tools to retrieve and represent multimedia information. In this study we designed and implemented a hypermedia system which has the hierarchical group and local map for effective navigation. We also propose the clustering mechanism which constructs a cluster tree and uses this knowledge for navigation. The system has been designed to integrate the browsing and searching function of the hypermedia system for efficient multimedia information retrieval and user-interface. This system can be used to develop hypermedia application systems in the area of encyclopedia, reference document information, electronic dictionary and electronic book.

  • PDF

A Multi-Strategic Mapping Approach for Distributed Topic Maps (분산 토픽맵의 다중 전략 매핑 기법)

  • Kim Jung-Min;Shin Hyo-phil;Kim Hyoung-Joo
    • Journal of KIISE:Software and Applications
    • /
    • v.33 no.1
    • /
    • pp.114-129
    • /
    • 2006
  • Ontology mapping is the task of finding semantic correspondences between two ontologies. In order to improve the effectiveness of ontology mapping, we need to consider the characteristics and constraints of data models used for implementing ontologies. Earlier research on ontology mapping, however, has proven to be inefficient because the approach should transform input ontologies into graphs and take into account all the nodes and edges of the graphs, which ended up requiring a great amount of processing time. In this paper, we propose a multi-strategic mapping approach to find correspondences between ontologies based on the syntactic or semantic characteristics and constraints of the topic maps. Our multi-strategic mapping approach includes a topic name-based mapping, a topic property-based mapping, a hierarchy-based mapping, and an association-based mapping approach. And it also uses a hybrid method in which a combined similarity is derived from the results of individual mapping approaches. In addition, we don't need to generate a cross-pair of all topics from the ontologies because unmatched pairs of topics can be removed by characteristics and constraints of the topic maps. For our experiments, we used oriental philosophy ontologies, western philosophy ontologies, Yahoo western philosophy dictionary, and Yahoo german literature dictionary as input ontologies. Our experiments show that the automatically generated mapping results conform to the outputs generated manually by domain experts, which is very promising for further work.

Contents and Organization of Geographic Education in Social Studies Curriculum-A Comparative Case Study of U.S. and Korea (한.미 지리교육의 내용과 조직 비교)

  • 박선미
    • Journal of the Korean Geographical Society
    • /
    • v.36 no.2
    • /
    • pp.191-210
    • /
    • 2001
  • 본 연구는 지리교육의 내용 선정과 조직에 대한 기초 연구로써, 우리나라 제 7차 사회과 규과정에서 명시하고 있는 지리교육의 방향, 내용 및 조직을 분석하고 뉴욕주의 2000년 에 제시된 그것과 비교하였다. 우리나라와 미국의 지리교육의 내용과 조직에 대한 연구는 양국의 교육과정 개발 관점에 대한 분석과 더불어 현대 사회에서 교과로써 경쟁력을 갖춘 지리교육의 방향 및 지리적 지식의 성격에 대한 논의를 바탕으로 한다. 뉴욕주의 지리교육은 <기본으로의 복귀> 분위기 속에서 지리교육에서 학습해야 할 6가지 핵심 요소를 추출하고 그에 따라 내용을 체계화한 반면 우리 나라의 경우 한국과 세계의 지역지리, 계통 지리의 내용을 나열적으로 학습하도록 하였다. 그러나 백과사전식 내용 구성은 정보화 사회에서 지리교육의 역할을 스스로 축소시키고 지리교육에 대한 학생의 관심과 흥미를 감소시키는 결과를 초래한다. 미국 뉴욕주의 지리 내용과 조직은 우리나라 지리교육과정의 내용 구성 방향에 많은 시사점을 줄 수 있다. 지리교육의 생장력을 키우기 위한 지리 내용은 현대 사회 및 학습자의 요구와 지리학의 경향과 쟁점에 근거하여 그 핵심 요소가 선정되어야 하고, 지역적 방법과 계통적 방법을 적절하게 결함시켜 역동적이고 흥미있게 조직되어야 한다.되어야 한다.

  • PDF

Analysis of Wikipedia Citations in Peer-Reviewed Journal Articles (학술논문에서의 위키피디아 인용에 관한 연구)

  • Shim, Wonsik;Byun, Jeayeon;Kim, Minjung
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.47 no.2
    • /
    • pp.247-264
    • /
    • 2013
  • Wikipedia represents a revolutionary experiment in the sense that it allows anonymous Internet users to contribute, change, and edit Encyclopedia contents used by tens of millions of people. While a very popular information source on the Internet, because of its questionable information credibility and accuracy, citing Wikipedia articles is being regarded as a risky behavior for scholars. The present study identified 282 scholarly articles from Thomson Reuters' Web of Science citation database that cite Wikipedia at least once. Out of the millions of articles indexed in Web of Science, the proportion of articles citing Wikipedia is extremely small. On the other hand, the numbers are showing a marked increase since 2011. Wikipedia citing articles are distributed in subject areas, such as library and information science, business, psychology, education, and communication more often than in other areas. The distribution of a total of 577 citations from 267 articles for which we were able to obtain full texts shows that Wikipedia is being cited mainly in studies of Wikipedia (139 citations, 24.1%) or as a ready reference source (331 citations, 57.4%). At the same time, about 15% of total citations turned out to be cases of potentially risky behaviors in which Wikipedia is being cited as a crucial basis or data source for study.

Study on the Representation Modes and Reality of Web Documentaries (웹다큐멘터리의 재현양식과 리얼리티에 관한 연구)

  • Jeon, Gyongran
    • Cartoon and Animation Studies
    • /
    • s.45
    • /
    • pp.259-282
    • /
    • 2016
  • Documentaries are being recreated into a new genre and the Web Documentary is the typical case. Web Documentaries are the documentaries those comprise creators and users and they are the novel type of text that the interaction with users is absolute. In this research, two Web Documentaries and are analyzed for examining how Web utilizes its features as expressive media inducing users to experience reality. Web Documentaries have dual and spatial structure that allows user interaction and make users to face with various information and knowledge about reality by its encyclopedic characteristics. Also, Web Documentaries give the role of progressing documentary and expanding text to users and that is, they stimulate users' consciousness reminding that they are the ones who explore through reality. In this process, users of Web Documentaries get potentiality of critically examining the reality suggested by documentaries and grasping the meanings beneath it. These features make Web Documentaries special contrast to traditional documentaries not only with their way of pursuing the reality but also with their meanings. This makes the innovative position of Web Documentaries phenomenon clear, issuing the necessity of the discussion about Web Documentaries more strongly. Web Documentaries are not just new media technological phenomenon, and they have their significance as a fundamental challenge toward traditional documentaries.

Development of Information Extraction System from Multi Source Unstructured Documents for Knowledge Base Expansion (지식베이스 확장을 위한 멀티소스 비정형 문서에서의 정보 추출 시스템의 개발)

  • Choi, Hyunseung;Kim, Mintae;Kim, Wooju;Shin, Dongwook;Lee, Yong Hun
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.111-136
    • /
    • 2018
  • In this paper, we propose a methodology to extract answer information about queries from various types of unstructured documents collected from multi-sources existing on web in order to expand knowledge base. The proposed methodology is divided into the following steps. 1) Collect relevant documents from Wikipedia, Naver encyclopedia, and Naver news sources for "subject-predicate" separated queries and classify the proper documents. 2) Determine whether the sentence is suitable for extracting information and derive the confidence. 3) Based on the predicate feature, extract the information in the proper sentence and derive the overall confidence of the information extraction result. In order to evaluate the performance of the information extraction system, we selected 400 queries from the artificial intelligence speaker of SK-Telecom. Compared with the baseline model, it is confirmed that it shows higher performance index than the existing model. The contribution of this study is that we develop a sequence tagging model based on bi-directional LSTM-CRF using the predicate feature of the query, with this we developed a robust model that can maintain high recall performance even in various types of unstructured documents collected from multiple sources. The problem of information extraction for knowledge base extension should take into account heterogeneous characteristics of source-specific document types. The proposed methodology proved to extract information effectively from various types of unstructured documents compared to the baseline model. There is a limitation in previous research that the performance is poor when extracting information about the document type that is different from the training data. In addition, this study can prevent unnecessary information extraction attempts from the documents that do not include the answer information through the process for predicting the suitability of information extraction of documents and sentences before the information extraction step. It is meaningful that we provided a method that precision performance can be maintained even in actual web environment. The information extraction problem for the knowledge base expansion has the characteristic that it can not guarantee whether the document includes the correct answer because it is aimed at the unstructured document existing in the real web. When the question answering is performed on a real web, previous machine reading comprehension studies has a limitation that it shows a low level of precision because it frequently attempts to extract an answer even in a document in which there is no correct answer. The policy that predicts the suitability of document and sentence information extraction is meaningful in that it contributes to maintaining the performance of information extraction even in real web environment. The limitations of this study and future research directions are as follows. First, it is a problem related to data preprocessing. In this study, the unit of knowledge extraction is classified through the morphological analysis based on the open source Konlpy python package, and the information extraction result can be improperly performed because morphological analysis is not performed properly. To enhance the performance of information extraction results, it is necessary to develop an advanced morpheme analyzer. Second, it is a problem of entity ambiguity. The information extraction system of this study can not distinguish the same name that has different intention. If several people with the same name appear in the news, the system may not extract information about the intended query. In future research, it is necessary to take measures to identify the person with the same name. Third, it is a problem of evaluation query data. In this study, we selected 400 of user queries collected from SK Telecom 's interactive artificial intelligent speaker to evaluate the performance of the information extraction system. n this study, we developed evaluation data set using 800 documents (400 questions * 7 articles per question (1 Wikipedia, 3 Naver encyclopedia, 3 Naver news) by judging whether a correct answer is included or not. To ensure the external validity of the study, it is desirable to use more queries to determine the performance of the system. This is a costly activity that must be done manually. Future research needs to evaluate the system for more queries. It is also necessary to develop a Korean benchmark data set of information extraction system for queries from multi-source web documents to build an environment that can evaluate the results more objectively.