• Title/Summary/Keyword: Data & Knowledge Engineering

Search Result 1,311, Processing Time 0.027 seconds

Building an Ontology for Structured Data Entry of Signs and Symptoms in Oriental Medicine (Protege를 이용한 한의학의 구조화된 증상 입력을 위한 온톨로지 개발)

  • Park Kyung Mo;Lim Hee Sook;Park Jong Hyun
    • Journal of Physiology & Pathology in Korean Medicine
    • /
    • v.17 no.5
    • /
    • pp.1151-1156
    • /
    • 2003
  • To obtain both of the fast and complete data entry and the acquisition of reusable data in a Computer-based Patient Record system (CPR), we are building the ontology that is used by the entry supporting agents. Our application domain is Traditional Chinese Medicine. As the tool for the implementation, we used protege 2000 which is ontology building tool and provides frame knowledge representation language. In this paper, the construction methodology of our ontology is reported.

Fault Detection, Diagnosis, and Optimization of Wafer Manufacturing Processes utilizing Knowledge Creation

  • Bae Hyeon;Kim Sung-Shin;Woo Kwang-Bang;May Gary S.;Lee Duk-Kwon
    • International Journal of Control, Automation, and Systems
    • /
    • v.4 no.3
    • /
    • pp.372-381
    • /
    • 2006
  • The purpose of this study was to develop a process management system to manage ingot fabrication and improve ingot quality. The ingot is the first manufactured material of wafers. Trace parameters were collected on-line but measurement parameters were measured by sampling inspection. The quality parameters were applied to evaluate the quality. Therefore, preprocessing was necessary to extract useful information from the quality data. First, statistical methods were used for data generation. Then, modeling was performed, using the generated data, to improve the performance of the models. The function of the models is to predict the quality corresponding to control parameters. Secondly, rule extraction was performed to find the relation between the production quality and control conditions. The extracted rules can give important information concerning how to handle the process correctly. The dynamic polynomial neural network (DPNN) and decision tree were applied for data modeling and rule extraction, respectively, from the ingot fabrication data.

A Study on Efficient Data De-Identification Method for Blockchain DID

  • Min, Youn-A
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.13 no.2
    • /
    • pp.60-66
    • /
    • 2021
  • Blockchain is a technology that enables trust-based consensus and verification based on a decentralized network. Distributed ID (DID) is based on a decentralized structure, and users have the right to manage their own ID. Recently, interest in self-sovereign identity authentication is increasing. In this paper, as a method for transparent and safe sovereignty management of data, among data pseudonymization techniques for blockchain use, various methods for data encryption processing are examined. The public key technique (homomorphic encryption) has high flexibility and security because different algorithms are applied to the entire sentence for encryption and decryption. As a result, the computational efficiency decreases. The hash function method (MD5) can maintain flexibility and is higher than the security-related two-way encryption method, but there is a threat of collision. Zero-knowledge proof is based on public key encryption based on a mutual proof method, and complex formulas are applied to processes such as personal identification, key distribution, and digital signature. It requires consensus and verification process, so the operation efficiency is lowered to the level of O (logeN) ~ O(N2). In this paper, data encryption processing for blockchain DID, based on zero-knowledge proof, was proposed and a one-way encryption method considering data use range and frequency of use was proposed. Based on the content presented in the thesis, it is possible to process corrected zero-knowledge proof and to process data efficiently.

A Study on The Development Methodology for Intelligent College Road Map Advice System (지능형 전공지도시스템 개발 방법론 연구)

  • Choi, Doug-Won;Cho, Kyung-Pil;Shin, Jin-Gyu
    • Journal of Intelligence and Information Systems
    • /
    • v.11 no.3
    • /
    • pp.57-67
    • /
    • 2005
  • Data mining techniques enable us to generate useful information for decision support from the data sources which are generated and accumulated in the process of routine organizational management activities. College administration system is a typical example that produces a warehouse of student records as each and every student enters a college and undertakes the curricular and extracurricular activities. So far, these data have been utilized to a very limited student service purposes, such as issuance of transcripts, graduation evaluation, GPA calculation, etc. In this paper, we utilized Holland career search test results, TOEIC score, course work list and GPA score as the input for data mining, and we were able to generate knowledge and rules with regard to the college road map advisory service. Factor analysis and AHP(Analytic Hierarchy Process) were the primary techniques deployed in the data mining process. Since these data mining techniques are very powerful in processing and discovering useful knowledge and information from large scale student databases, we can expect a highly sophisticated student advisory knowledge and services which may not be obtained from the human student advice experts.

  • PDF

The method of using database technology to process rules of Rule-Based System

  • Zheng, Baowei;Yeo, Jeong-Mo
    • Journal of information and communication convergence engineering
    • /
    • v.8 no.1
    • /
    • pp.89-94
    • /
    • 2010
  • The most important of rule-base system is the knowledge base that determines the power of rule-base system. The important form of this knowledge is how to descript kinds of rules. The Rule-Base System (RBS) has been using in many field that need reflect quickly change of business rules in management system. As far, when develop the Rule-Based System, we must make a rule engine with a general language. There are three disadvantage of in this developed method. First, while there are many data that must be processed in the system, the speed of processing data will become very slow so that we cannot accept it. Second, we cannot change the current system to make it adaptive to changes of business rules as quickly as possible. Third, large data make the rule engine become very complex. Therefore, in this paper, we propose the two important methods of raising efficiency of Rule-Base System. The first method refers to using the Relational database technology to process the rules of the Rule-Base System, the second method refers to a algorithm of according to Quine McCluskey formula compress the rows of rule table. Because the expressive languages of rule are still remaining many problems, we will introduce a new expressive language, which is Rule-Base Data Model short as RBDM in this paper.

Splitting Algorithm Using Total Information Gain for a Market Segmentation Problem

  • Kim, Jae-Kyeong;Kim, Chang-Kwon;Kim, Soung-Hie
    • Journal of the Korean Operations Research and Management Science Society
    • /
    • v.18 no.2
    • /
    • pp.183-203
    • /
    • 1993
  • One of the most difficult and time-consuming stages in the development of the knowledge-based system is a knowledge acquisition. A splitting algorithm is developed to infer a rule-tree which can be converted to a rule-typed knowledge. A market segmentation may be performed in order to establish market strategy suitable to each market segment. As the sales data of a product market is probabilistic and noisy, it becomes necessary to prune the rule-tree-at an acceptable level while generating a rule-tree. A splitting algorithm is developed using the pruning measure based on a total amount of information gain and the measure of existing algorithms. A user can easily adjust the size of the resulting rule-tree according to his(her) preferences and problem domains. The algorithm is applied to a market segmentation problem of a medium-large computer market. The algorithm is illustrated step by step with a sales data of a computer market and is analyzed.

  • PDF

Electrical Fire Cause Diagnosis System based on Fuzzy Inference

  • Lee, Jong-Ho;Kim, Doo-Hyun
    • International Journal of Safety
    • /
    • v.4 no.2
    • /
    • pp.12-17
    • /
    • 2005
  • This paper aims at the development of an knowledge base for an electrical fire cause diagnosis system using the entity relation database. The relation database which provides a very simple but powerful way of representing data is widely used. The system focused on database construction and cause diagnosis can diagnose the causes of electrical fires easily and efficiently. In order to store and access to the information concerned with electrical fires, the key index items which identify electrical fires uniquely are derived out. The knowledge base consists of a case base which contains information from the past fires and a rule base with rules from expertise. To implement the knowledge base, Access 2000, one of DB development tools under windows environment and Visual Basic 6.0 are used as a DB building tool. For the reasoning technique, a mixed reasoning approach of a case based inference and a rule based inference has been adopted. Knowledge-based reasoning could present the cause of a newly occurred fire to be diagnosed by searching the knowledge base for reasonable matching. The knowledge-based database has not only searching functions with multiple attributes by using the collected various information(such as fire evidence, structure, and weather of a fire scene), but also more improved diagnosis functions which can be easily wed for the electrical fire cause diagnosis system.

Compression Conversion and Storing of Large RDF datasets based on MapReduce (맵리듀스 기반 대량 RDF 데이터셋 압축 변환 및 저장 방법)

  • Kim, InA;Lee, Kyong-Ha;Lee, Kyu-Chul
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.4
    • /
    • pp.487-494
    • /
    • 2022
  • With the recent demand for analysis using data, the size of the knowledge graph, which is the data to be analyzed, gradually increased, reaching about 82 billion edges when extracted from the web as a knowledge graph. A lot of knowledge graphs are represented in the form of Resource Description Framework (RDF), which is a standard of W3C for representing metadata for web resources. Because of the characteristics of RDF, existing RDF storages have the limitations of processing time overhead when converting and storing large amounts of RDF data. To resolve these limitations, in this paper, we propose a method of compressing and converting large amounts of RDF data into integer IDs using MapReduce, and vertically partitioning and storing them. Our proposed method demonstrated a high performance improvement of up to 25.2 times compared to RDF-3X and up to 3.7 times compared to H2RDF+.

Consumers' Attitudes toward the General and Fashion-Specific Climate Environments: Focusing on the Relations with Values, Knowledge, and Climate Cognition (소비자들의 일반기후환경태도와 패션기후환경태도: 가치와 지식 및 기후인식과의 관계를 중심으로)

  • Ihn Hee Chung
    • Human Ecology Research
    • /
    • v.61 no.4
    • /
    • pp.599-613
    • /
    • 2023
  • This study investigated female consumers' attitudes toward the general and fashion-specific climate environments and analyzed the relations between the attitudes and the variables such as values, knowledge, and climate cognition. The data was collected from a sample of 450 women in their 20s, 30s, and 40s via quota sampling from a selfreported online survey in 2023. The measurement comprised the attitudes toward the general and fashion-specific climate environments, Rokeach's 18 terminal values, Holbrook's 8 consumer values regarding fashion products, climate environmental knowledge related to fashion, the cognition concerning the climate crisis, and several demographic variables. Descriptive statistics, factor analysis, reliability analysis, and correlations were applied to the data using SPSS. As a result, two factors were determined for the attitudes toward the general and fashion-specific climate environments, respectively: social and personal. Family security, happiness, and self-respect were identified as important terminal values. Quality, efficiency, aesthetics, and ethics were considered important when the current sample group purchased fashion products. The mean score of climate environmental knowledge related to fashion was lower than neutral; however the cognition of the climate crisis was considerably high. Attitudes toward the general and fashion-specific climate environments showed positive relations with values, knowledge, and climate cognition. The results were discussed to provide some insight and suggestions to carbon neutrality and the related studies.

A Study on the Development of Computer Aided Die Design System for Lead Frame, Semiconductor (반도체 리드 프레임의 금형설계 자동화 시스템 개발에 관한 연구)

  • Choe, Jae-Chan;Kim, Byeong-Min;Kim, Cheol;Kim, Jae-Hun;Kim, Chang-Bong
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.16 no.6
    • /
    • pp.123-132
    • /
    • 1999
  • This paper describes a research work of developing computer-aided design of lead frame, semiconductor, with blanking operation which is very precise for progressive working. Approach to the system is based on the knowledge-based rules. Knowledge for the system is formulated from pasticity theories, experimental results and the empirical knowledge of field experts. This system has been written in AutoLISP on the AutoCAD using a personal computer and in I-DEAS Drafting Programming Language on the I-DEAS Master Series Drafting with Workstation, HP9000/715(64). Transference of data between AutoCAD and I-DEAS Master Series Drafting is accomplished by DXF(drawing exchange format) and IGES(initial graphics exchange specification) methods. This system is composed of five modules, which are input and shape treatment, production feasibility check, strip-layout, data-conversion and die-layout modules. The process planning and Die design system is designed by considering several factors, such as complexities of blank geometry, punch profiles, and the availability of a press equipment and standard parts. This system provides its efficiecy for strip-layout, and die design for lead frame, semiconductor.

  • PDF