• Title/Summary/Keyword: 데이타 변환

Search Result 305, Processing Time 0.024 seconds

Neural Network Refinement using Hidden Knowledge Extraction (은닉지식 추출을 이용한 신경망회로망 정제)

  • Kim, Hyeon-Cheol
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.11
    • /
    • pp.1082-1087
    • /
    • 2000
  • 신경회로망 구조의 정제(精製)는 회로망의 일반화능력이나 효율성의 관점에서 중요한 문제이다. 본 논문에서는 feed-forward neural networks로부터 은닉지식을 추출하는 방법을 사용하여 네트워크 재구성을 통한 정제방법을 제안한다. 먼저, 효율적인 if-then rule 추출방법을 제시하고 그 추출된 룰들을 사용하여 룰기반 네트워크로 변환하는 과정을 보여준다. 생성된 룰기반 네트워크 fully connected network에 비하여 상당히 축소된 연결 복잡도를 가지게 되며 일반적으로 더 우수한 일반화능력을 가지게 된다. 본 연구는 도메인 지식이 없이 데이타만 사용하여 어떻게 정제된 룰기반 신경망회로를 생성하고 있는가를 보여준다. 도메인 데이타들에 대한 실험결과도 제시하였다.

  • PDF

Implementation of CORBA based Spatial Data Provider for Interoperability (상호운용을 지원하는 코바 기반 공간 데이터 제공자의 설계 및 구현)

  • Kim, Min-Seok;An, Kyoung-Hwan;Hong, Bong-Hee
    • Journal of Korea Spatial Information System Society
    • /
    • v.1 no.2 s.2
    • /
    • pp.33-46
    • /
    • 1999
  • In distributed computing platforms like CORBA, wrappers are used to integrate heterogeneous systems or databases. A spatial data provider is one of the wrappers because it provides clients with uniform access interfaces to diverse data sources. The individual implementation of spatial data providers for each of different data sources is not efficient because of redundant coding of the wrapper modules. This paper presents a new architecture of the spatial data provider which consists of two layered objects : independent wrapper components and dependent wrapper components. Independent wrapper components would be reused for implementing a new data provider for a new data source, which dependent wrapper components should be newly coded for every data source. This paper furthermore discussed the issues of implementing the representation of query results in the middleware. There are two methods of keeping query results in the middleware. One is to keep query results as non-CORBA objects and the other is to transform query results into CORBA objects. The evaluation of the above two methods shows that the cost of making CORBA objects is very expensive.

  • PDF

Design of The Geographic Information Database Structure for Census Mapping (센서스 지도제작을 위한 지리정보데이타베이스 구조연구)

  • 김설희
    • Spatial Information Research
    • /
    • v.1 no.1
    • /
    • pp.17-28
    • /
    • 1993
  • In order to minimize vectorizing tasks, which require huge reso¬urces and time and to support the census mapping effectively, the geographic information databases structure has been studied. The steps of the new approach are as follows. : Step 1, Scanning the maps of the whole country and storing the image data in raster format. Step 2, Vectorizing the data of specific items for Census operation such as Enume¬ration District, and then linking to attribute data in the text format. Step 3, Designing the database with a Tile and Multi-layer structure to make a continuous map logically. Step 4, Implement Censlls Mapping System(CMS) for efficient mapping and retrieving. As a consequence of this study, the cost, manpower and time effectiveness was proved and it was confirmed to produce lIseful and high-qual ified maps for the Census. In the future, this system wi II be able to provide many organizations and individuals with the various data based on geographical statistical information.

  • PDF

A System Model for Storage Independent Use of SPARQL-to-SQL Translation Algorithm (SPARQL-to-SQL 변환 알고리즘의 저장소 독립적 활용을 위한 시스템 모델)

  • Son, Ji-Seong;Jeong, Dong-Won;Baik, Doo-Kwon
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.5
    • /
    • pp.467-471
    • /
    • 2008
  • With active research on Web ontology, various storages and query languages have been developed to store Web Ontology. As SPARQL usage increases and most of storages are based on relational data base, the necessity of SPARQL-to-SQL translation algorithm development becomes issued. Even though several translation algorithms have been proposed, there still remain the following problems. They do not support fully SPARQL clauses and they are dependent on a specific storage model. This paper proposes a new model to use a specific translation algorithm independently on storages.

NBR-Safe Transform: Lower-Dimensional Transformation of High-Dimensional MBRs in Similar Sequence Matching (MBR-Safe 변환 : 유사 시퀀스 매칭에서 고차원 MBR의 저차원 변환)

  • Moon, Yang-Sae
    • Journal of KIISE:Databases
    • /
    • v.33 no.7
    • /
    • pp.693-707
    • /
    • 2006
  • To improve performance using a multidimensional index in similar sequence matching, we transform a high-dimensional sequence to a low-dimensional sequence, and then construct a low-dimensional MBR that contains multiple transformed sequences. In this paper we propose a formal method that transforms a high-dimensional MBR itself to a low-dimensional MBR, and show that this method significantly reduces the number of lower-dimensional transformations. To achieve this goal, we first formally define the new notion of MBR-safe. We say that a transform is MBR-safe if a low-dimensional MBR to which a high-dimensional MBR is transformed by the transform contains every individual low-dimensional sequence to which a high-dimensional sequence is transformed. We then propose two MBR-safe transforms based on DFT and DCT, the most representative lower-dimensional transformations. For this, we prove the traditional DFT and DCT are not MBR-safe, and define new transforms, called mbrDFT and mbrDCT, by extending DFT and DCT, respectively. We also formally prove these mbrDFT and mbrDCT are MBR-safe. Moreover, we show that mbrDFT(or mbrDCT) is optimal among the DFT-based(or DCT-based) MBR-safe transforms that directly convert a high-dimensional MBR itself into a low-dimensional MBR. Analytical and experimental results show that the proposed mbrDFT and mbrDCT reduce the number of lower-dimensional transformations drastically, and improve performance significantly compared with the $na\"{\i}ve$ transforms. These results indicate that our MBR- safe transforms provides a useful framework for a variety of applications that require the lower-dimensional transformation of high-dimensional MBRs.

Applying In-Page Logging to SQLite DBMS (SQLite DBMS에 IPL 기법 응용)

  • Na, Gap-Joo;Kim, Sang-Woo;Kim, Jae-Myung;Lee, Sang-Won
    • Journal of KIISE:Databases
    • /
    • v.35 no.5
    • /
    • pp.400-410
    • /
    • 2008
  • Flash memory has been widely used in mobile devices, such as mobile phone and digital camera. Recently flash SSD(Solid State Disk), having same interface of the disk drive, is replacing the hard disk of some laptop computers. However, flash memory still cannot be considered as the storage of database systems. The FTL(Flash Translation Layer) of commercial flash SSD, making flash memory operate exactly same as a hard disk, shows poor performance on the workload of databases with many random overwrites. Recently In-Page Logging(IPL) approach was proposed to solve this problem. In this paper, we implement IPL approach on SQLite, a popular open source embedded DBMS, and evaluate its performance. It improves the performance by up to 30 factors for update queries.

High-Dimensional Clustering Technique using Incremental Projection (점진적 프로젝션을 이용한 고차원 글러스터링 기법)

  • Lee, Hye-Myung;Park, Young-Bae
    • Journal of KIISE:Databases
    • /
    • v.28 no.4
    • /
    • pp.568-576
    • /
    • 2001
  • Most of clustering algorithms data to degenerate rapidly on high dimensional spaces. Moreover, high dimensional data often contain a significant a significant of noise. which causes additional ineffectiveness of algorithms. Therefore it is necessary to develop algorithms adapted to the structure and characteristics of the high dimensional data. In this paper, we propose a clustering algorithms CLIP using the projection The CLIP is designed to overcome efficiency and/or effectiveness problems on high dimensional clustering and it is the is based on clustering on each one dimensional subspace but we use the incremental projection to recover high dimensional cluster and to reduce the computational cost significantly at time To evaluate the performance of CLIP we demonstrate is efficiency and effectiveness through a series of experiments on synthetic data sets.

  • PDF

Query Rewriting and Indexing Schemes for Distributed Systems based on the Semantic Web (시맨틱 웹 기반의 분산 시스템을 위한 질의 변환 및 인덱싱 기법)

  • Chae, Kwang-Ju;Kim, Youn-Hee;Lim, Hae-Chull
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.7
    • /
    • pp.718-722
    • /
    • 2008
  • Ontology plays an important role of the Semantic Web to describe meaning and reasoning of resources. Ontology has more rich expressive power through OWL that is a next standard representation language recommended by W3C. As the Semantic Web is widely known, an amount of information resources on the Web is growing rapidly and the related information resources are placed in distributed systems on the Web. So, for providing seamless services without the awareness of far distance, efficient management of the distributed information resources is required. Especially, sear ching fast for local repositories that include data related to user's queries is important to the performance of systems in the distributed environment. In this paper, first, we propose an index structure to search local repositories related to queries in the distributed Semantic Web. Second, we propose a query rewriting strategy to extend given user's query using various expression of OWL. Through the proposed index and query strategy, we can utilize various expressions of OWL and find local repositories related to all query patterns on the Semantic Web.

An XML-QL to SQL Translator for Processing XML Data (XML 데이타 처리를 위한 XML-QL to SQL 번역기)

  • Jang, Gyeong-Ja;Lee, Gi-Ho
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.1
    • /
    • pp.1-8
    • /
    • 2002
  • XML has been proposed as an international standard for organizing and exchanging a great diversity of the Web data. It is important to retrieve components of stored XML documents that are needed by a wide variety of applications. In this paper, we suggest a method to store XML documents and to retrieve an XML data. In other words, we suggest the method of retrieving XML data is using XML -QL. So we need to mapping XML-QL to SQL translator on top of an RDBMS. The contributions of this paper include, besides the detailed design and implementation of the translator, demonstration of feasibility of such a translator, and a comprehensive classification of XML queries and their mappings to SQL relational queries.

인터넷 질의 처리를 위한 웨이블릿 변환에 기반한 통합 요약정보의 관리

  • Joe, Moon-Jeung;Whang, Kyu-Young;Kim, Sang-Wook;Shim, Kyu-Seok
    • Journal of KIISE:Databases
    • /
    • v.28 no.4
    • /
    • pp.702-714
    • /
    • 2001
  • As Internet technology evolves, there is growing need of Internet queries involving multiple information sources. Efficient processing of such queries necessitates the integrated summary data that compactly represents the data distribution of the entire database scattered over many information sources. This paper presents an efficient method of managing the integrated summary data based on the wavelet transform and addresses Internet query processing using the integrated summary data. The simplest method for creating the integrated summary data would be to summarize the integrated data sidtribution obtained by merging the data distributions in multiple information sources. However, this method suffers from the high cost of transmitting storing and merging a large amount of data distribution. To overcome the drawbacks, we propose a new wavelet transform based method that creates the integrated summary data by merging multiple summary data and effective method for optimizing Internet queries using it A wavelet transformed summary data is converted to satisfy conditions for merging. Moreover i the merging process is very simpe owing to the properties of the wavelet transform. we formally derive the upper bound of the error of the wavelet transformed intergrated summary data. Compared with the histogram-based integrated summary data the wavelet transformedintegrated summary data provesto be 1.6~5.5 time more accurate when used for selectivity estimation in experiments. In processing Internet top-N queries involving 56 information sources using the integrated summary data reduces the processing cost to 1/44 of the cost of not using it.

  • PDF