• Title/Summary/Keyword: 트리플 추출

Search Result 39, Processing Time 0.021 seconds

Reference Resolution for Ontology Population (온톨로지 인스턴스 생성을 위한 상호참조 해결 연구)

  • Choi, Miran;Lee, Changki;Wang, Jihyun;Jang, Muyng-Gil
    • Annual Conference on Human and Language Technology
    • /
    • 2007.10a
    • /
    • pp.140-144
    • /
    • 2007
  • 시맨틱 웹 기술의 주축을 이루는 온톨로지의 구축시에 인스턴스를 생성하기 위하여 대상 문서를 구성하는 자연어 문장을 텍스트 마이닝 기술을 이용하여 트리플을 추출한다. 인스턴스를 생성할 때 보다 많은 정보를 추출하기 위해서 문장에 나타나는 상호참조 해결이 필요하다. 본 연구에서는 문서에서 많이 나타나는 명사구로 이루어진 대용어를 해석하기 위하여 언어 분석된 다양한 결과 정보를 이용한다. 본 연구에서는 계층적인 의미구조와 청킹을 이용한 규칙기반의 상호참조 해결 방법을 제안하고 실험을 통해 알고리즘의 정확도를 제시한다.

  • PDF

Con-Talky: Information Extraction and Visualization Platform for Communication of Construction Industry (Con-Talky: 건설 분야 전문가의 의사소통을 위한 정보 추출 및 시각화 플랫폼)

  • Shim, Midan;Park, Chanjun;Hur, Yuna;Lim, Heuiseok
    • Annual Conference on Human and Language Technology
    • /
    • 2021.10a
    • /
    • pp.476-481
    • /
    • 2021
  • 본 논문은 용어의 비통일성과 문서의 다양성으로 인해 발생하는 건설분야 전문가들의 의사소통 문제를 해결하기 위한 Con-Talky를 제안한다. Con-Talky는 자연언어처리의 대표적인 기술인 형태소분석, 의존구문분석, 의미역 결정 기술을 융합하여 건설분야의 "설계기준문서"를 시각화하고 핵심 정보추출을 자동으로 해주는 플랫폼이다. 해당 플랫폼을 이용하여 토목분야 전문가들의 의사소통 문제를 완화시킬 수 있으며 용어의 비통일성 및 표준화에도 기여할 수 있다. 또한 본 논문은 국내 건설 및 토목분야에 최초로 자연언어처리 기술을 적용한 논문이다. 해당 분야의 연구를 활성화 하기 위해 건설분야에 특화된 단일 말뭉치와 트리플 데이터를 자체 제작함과 동시에 전면 공개하였다.

  • PDF

Integration of Extended IFC-BIM and Ontology for Information Management of Bridge Inspection (확장 IFC-BIM 기반 정보모델과 온톨로지를 활용한 교량 점검데이터 관리방법)

  • Erdene, Khuvilai;Kwon, Tae Ho;Lee, Sang-Ho
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.33 no.6
    • /
    • pp.411-417
    • /
    • 2020
  • To utilize building information modeling (BIM) technology at the bridge maintenance stage, it is necessary to integrate large quantities of bridge inspection and model data for object-oriented information management. This research aims to establish the benefits of utilizing the extended industry foundation class (IFC)-BIM and ontology for bridge inspection information management. The IFC entities were extended to represent the bridge objects, and a method of generating the extended IFC-based information model was proposed. The bridge inspection ontology was also developed by extraction and classification of inspection concepts from the AASHTO standard. The classified concepts and their relationships were mapped to the ontology based on the semantic triples approach. Finally, the extended IFC-based BIM model was integrated with the ontology for bridge inspection data management. The effectiveness of the proposed framework for bridge inspection information management by integration of the extended IFC-BIM and ontology was tested and verified by extracting bridge inspection data via the SPARQL query.

Study on the Automatic Generation of DB Schema in URI Server (URI 서버 내 DB 스키마 자동 생성 연구)

  • Jung, Hanmin;Lee, MiKyung;Kang, In-Su;Sung, Won-Kyung
    • Annual Conference on Human and Language Technology
    • /
    • 2007.10a
    • /
    • pp.107-113
    • /
    • 2007
  • 본 연구는 시맨틱 데이터 정합성 검사와 RDF 트리플 생성 기능을 가진 URI 서버 내 DB 스키마의 자동 생성 방안을 기술한다. URI 서버는 시맨틱 웹 기술 기반 프레임워크 구성에 필요한 핵심 엔진으로서 인스턴스 생성 관리기, 검색 엔진, 추론 엔진 등과의 상호 작용을 통해 시맨틱 데이터를 서비스한다. 도메인이 바뀌거나 서비스가 변경되는 경우 DB 스키마를 전문가에 의해 수작업으로 생성함으로써 서비스 환경 변화에 즉각 대응하지 못하는 문제점을 가진 기존 URI 서버를, 적재되는 온톨로지에 따라 DB 스키마를 동적으로 자동 생성하는 방식으로 변경함으로써 다양한 응용 분야와 도메인에의 높은 이식성(Portability)을 보장할 수 있도록 한다. 기반정보 온톨로지, 응용 온톨로지, 개인화 온톨로지 등 3개 온톨로지 스키마와 11만 건 이상의 Citeseer Open Access Metadata로부터 추출된 인스턴스를 대상으로 한 실험을 통해 URI 서버 내 DB 스키마 자동 생성 가능성을 실증하였다.

  • PDF

User Interaction-based Graph Query Formulation and Processing (사용자 상호작용에 기반한 그래프질의 생성 및 처리)

  • Jung, Sung-Jae;Kim, Taehong;Lee, Seungwoo;Lee, Hwasik;Jung, Hanmin
    • Journal of KIISE:Databases
    • /
    • v.41 no.4
    • /
    • pp.242-248
    • /
    • 2014
  • With the rapidly growing amount of information represented in RDF format, efficient querying of RDF graph has become a fundamental challenge. SPARQL is one of the most widely used query languages for retrieving information from RDF dataset. SPARQL is not only simple in its syntax but also powerful in representation of graph pattern queries. However, users need to make a lot of efforts to understand the ontology schema of a dataset in order to compose a relevant SPARQL query. In this paper, we propose a graph query formulation and processing scheme based on ontology schema information which can be obtained by summarizing RDF graph. In the context of the proposed querying scheme, a user can interactively formulate the graph queries on the graphic user interface without making efforts to understand the ontology schema and even without learning SPARQL syntax. The graph query formulated by a user is transformed into a set of class paths, which are stored in a relational database and used as the constraint for search space reduction when the relational database executes the graph search operation. By executing the LUBM query 2, 8, and 9 over LUBM (10,0), it is shown that the proposed querying scheme returns the complete result set.

Pattern and Instance Generation for Self-knowledge Learning in Korean (한국어 자가 지식 학습을 위한 패턴 및 인스턴스 생성)

  • Yoon, Hee-Geun;Park, Seong-Bae
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.25 no.1
    • /
    • pp.63-69
    • /
    • 2015
  • There are various researches which proposed an automatic instance generation from freetext on the web. Existing researches that focused on English, adopts pattern representation which is generated by simple rules and regular expression. These simple patterns achieves high performance, but it is not suitable in Korean due to differences of characteristics between Korean and English. Thus, this paper proposes a novel method for generating patterns and instances which focuses on Korean. A proposed method generates high quality patterns by taking advantages of dependency relations in a target sentences. In addition, a proposed method overcome restrictions from high degree of freedom of word order in Korean by utilizing postposition and it identifies a subject and an object more reliably. In experiment results, a proposed method shows higher precision than baseline and it is implies that proposed approache is suitable for self-knowledge learning system.

Analysis of Soil Samples Obtained from Piston Sampler and Large Diameter Sampler (피스톤 샘플러와 대구경 샘플러를 이용한 시료 샘플의 공학적 분석)

  • Kim, Young Chin;Kang, Jae Mo
    • Journal of the Korean GEO-environmental Society
    • /
    • v.9 no.3
    • /
    • pp.29-34
    • /
    • 2008
  • A large diameter sampler was developed to take undisturbed samples from not only soft ground but also sandy and weathered ground. The large diameter sampler which was developed in Korea Institute of Construction Technology(KICT-type large diameter sampler) was manufactured based on the principle of triple core barrel sampling. A specially designed cutting device was used to cut and contain various kinds of samples in the sampler during a sampling and retrieval procedure. By adjusting the stiffness of the spring located at the top of the sampler, the distance between the cutting shoe and auger can be controlled in accordance with the ground condition. In order to investigate the applicability of the developed sampler and compare the quality of the samples taken by the sampler with that by the traditional thin-walled tube sampler, samples were taken at various sites according to the ground condition. And a series of laboratory tests such as the unconfined compress ion test, triaxial compression test, oedometer test, large diameter Rowe cell consolidation test (D: 150 mm) were performed. The test results showed that the samples by the KICT-type large diameter sampler show higher quality than the samples by the thin-walled tube sampler. And the validity and applicability of the developed KICT-type large diameter sampler was confirmed accordingly.

  • PDF

Implementation of Ontology-based Service by Exploiting Massive Crime Investigation Records: Focusing on Intrusion Theft (대규모 범죄 수사기록을 활용한 온톨로지 기반 서비스 구현 - 침입 절도 범죄 분야를 중심으로 -)

  • Ko, Gun-Woo;Kim, Seon-Wu;Park, Sung-Jin;No, Yoon-Joo;Choi, Sung-Pil
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.53 no.1
    • /
    • pp.57-81
    • /
    • 2019
  • An ontology is a complex structure dictionary that defines the relationship between terms and terms related to specific knowledge in a particular field. There have been attempts to construct various ontologies in Korea and abroad, but there has not been a case in which a large scale crime investigation record is constructed as an ontology and a service is implemented through the ontology. Therefore, this paper describes the process of constructing an ontology based on information extracted from instrusion theft field of unstructured data, a crime investigation document, and implementing an ontology-based search service and a crime spot recommendation service. In order to understand the performance of the search service, we have tested Top-K accuracy measurement, which is one of the accuracy measurement methods for event search, and obtained a maximum accuracy of 93.52% for the experimental data set. In addition, we have obtained a suitable clue field combination for the entire experimental data set, and we can calibrate the field location information in the database with the performance of F1-measure 76.19% Respectively.

A Study on Ontology and Topic Modeling-based Multi-dimensional Knowledge Map Services (온톨로지와 토픽모델링 기반 다차원 연계 지식맵 서비스 연구)

  • Jeong, Hanjo
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.79-92
    • /
    • 2015
  • Knowledge map is widely used to represent knowledge in many domains. This paper presents a method of integrating the national R&D data and assists of users to navigate the integrated data via using a knowledge map service. The knowledge map service is built by using a lightweight ontology and a topic modeling method. The national R&D data is integrated with the research project as its center, i.e., the other R&D data such as research papers, patents, and reports are connected with the research project as its outputs. The lightweight ontology is used to represent the simple relationships between the integrated data such as project-outputs relationships, document-author relationships, and document-topic relationships. Knowledge map enables us to infer further relationships such as co-author and co-topic relationships. To extract the relationships between the integrated data, a Relational Data-to-Triples transformer is implemented. Also, a topic modeling approach is introduced to extract the document-topic relationships. A triple store is used to manage and process the ontology data while preserving the network characteristics of knowledge map service. Knowledge map can be divided into two types: one is a knowledge map used in the area of knowledge management to store, manage and process the organizations' data as knowledge, the other is a knowledge map for analyzing and representing knowledge extracted from the science & technology documents. This research focuses on the latter one. In this research, a knowledge map service is introduced for integrating the national R&D data obtained from National Digital Science Library (NDSL) and National Science & Technology Information Service (NTIS), which are two major repository and service of national R&D data servicing in Korea. A lightweight ontology is used to design and build a knowledge map. Using the lightweight ontology enables us to represent and process knowledge as a simple network and it fits in with the knowledge navigation and visualization characteristics of the knowledge map. The lightweight ontology is used to represent the entities and their relationships in the knowledge maps, and an ontology repository is created to store and process the ontology. In the ontologies, researchers are implicitly connected by the national R&D data as the author relationships and the performer relationships. A knowledge map for displaying researchers' network is created, and the researchers' network is created by the co-authoring relationships of the national R&D documents and the co-participation relationships of the national R&D projects. To sum up, a knowledge map-service system based on topic modeling and ontology is introduced for processing knowledge about the national R&D data such as research projects, papers, patent, project reports, and Global Trends Briefing (GTB) data. The system has goals 1) to integrate the national R&D data obtained from NDSL and NTIS, 2) to provide a semantic & topic based information search on the integrated data, and 3) to provide a knowledge map services based on the semantic analysis and knowledge processing. The S&T information such as research papers, research reports, patents and GTB are daily updated from NDSL, and the R&D projects information including their participants and output information are updated from the NTIS. The S&T information and the national R&D information are obtained and integrated to the integrated database. Knowledge base is constructed by transforming the relational data into triples referencing R&D ontology. In addition, a topic modeling method is employed to extract the relationships between the S&T documents and topic keyword/s representing the documents. The topic modeling approach enables us to extract the relationships and topic keyword/s based on the semantics, not based on the simple keyword/s. Lastly, we show an experiment on the construction of the integrated knowledge base using the lightweight ontology and topic modeling, and the knowledge map services created based on the knowledge base are also introduced.