• Title/Summary/Keyword: Information Database

Search Result 8,518, Processing Time 0.029 seconds

An Efficient Algorithm for Mining Frequent Sequences In Spatiotemporal Data

  • Vhan Vu Thi Hong;Chi Cheong-Hee;Ryu Keun-Ho
    • 한국공간정보시스템학회:학술대회논문집
    • /
    • 2005.11a
    • /
    • pp.61-66
    • /
    • 2005
  • Spatiotemporal data mining represents the confluence of several fields including spatiotemporal databases, machine loaming, statistics, geographic visualization, and information theory. Exploration of spatial data mining and temporal data mining has received much attention independently in knowledge discovery in databases and data mining research community. In this paper, we introduce an algorithm Max_MOP for discovering moving sequences in mobile environment. Max_MOP mines only maximal frequent moving patterns. We exploit the characteristic of the problem domain, which is the spatiotemporal proximity between activities, to partition the spatiotemporal space. The task of finding moving sequences is to consider all temporally ordered combination of associations, which requires an intensive computation. However, exploiting the spatiotemporal proximity characteristic makes this task more cornputationally feasible. Our proposed technique is applicable to location-based services such as traffic service, tourist service, and location-aware advertising service.

  • PDF

Buffer Invalidation Schemes for High Performance Transaction Processing in Shared Database Environment (공유 데이터베이스 환경에서 고성능 트랜잭션 처리를 위한 버퍼 무효화 기법)

  • 김신희;배정미;강병욱
    • The Journal of Information Systems
    • /
    • v.6 no.1
    • /
    • pp.159-180
    • /
    • 1997
  • Database sharing system(DBSS) refers to a system for high performance transaction processing. In DBSS, the processing nodes are locally coupled via a high speed network and share a common database at the disk level. Each node has a local memory, a separate copy of operating system, and a DBMS. To reduce the number of disk accesses, the node caches database pages in its local memory buffer. However, since multiple nodes may be simultaneously cached a page, cache consistency must be ensured so that every node can always access the latest version of pages. In this paper, we propose efficient buffer invalidation schemes in DBSS, where the database is logically partitioned using primary copy authority to reduce locking overhead. The proposed schemes can improve performance by reducing the disk access overhead and the message overhead due to maintaining cache consistency. Furthermore, they can show good performance when database workloads are varied dynamically.

  • PDF

An Optimal Database Design Considering Denormalization in Relational Database (관계형 데이터베이스에서 비정규화를 고려한 최적 데이터베이스 설계)

  • 장영관;강맹규
    • The Journal of Information Technology and Database
    • /
    • v.3 no.1
    • /
    • pp.3-24
    • /
    • 1996
  • Databases are critical to business information systems, and RDBMS is most widely used for the database system. Normalization has been designed to control various anomalies(insert, update, and delete anomalies). However, normalized database design does not account for the tradeoffs necessary for the performance. In this research, we develop a model for database design by denormalization of duplicating attributes in order to reduce frequent join processes. In this mood, we consider insert, update, delete, and query costs. The anomaly and data inconsistency are removed by additional disk I/O which is necessary for each update and insert transaction. We propose a branch and bound method for this model, and show considerable cost reduction.

  • PDF

Performance of Database Driven Network Applications from the User Perspective

  • Tang, Shanyu;YongFeng, Huang;Yip, Yau Jim
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.3 no.3
    • /
    • pp.235-250
    • /
    • 2009
  • An understanding of performance of database driven network applications is critical in combating slow performance of e-commerce web sites, besides efficient web page design, and high quality and well-maintained web server equipment. This paper describes a method of measuring performance from the user viewpoint, which can help enormously in making realistic assessment of true performance of database driven applications. The performance measurements were taken at user locations by using several specially designed JavaScript functions along with ASP scripts. A performance study is presented in this paper, comparing performance of data access using stored procedures with the traditional way of querying a database. It is generally believed that stored procedures have performance benefits as they are pre-compiled. However, our study shows that the data access approach using stored procedures provides significant benefits(by about 30%) over the traditional approach for querying a commercial MySQL database, only when retrieving a substantial amount of data(at least 10,000 rows of data).

Development of Practical Data Mining Methods for Database Summarization

  • Lee, Do-Heon
    • The Journal of Information Technology and Database
    • /
    • v.4 no.2
    • /
    • pp.33-45
    • /
    • 1998
  • Database summarization is the procedure to obtain generalized and representative descriptions expressing the content of a large amount of database at a glance. We present a top-down summary refinement procedure to discover database summaries. The procedure exploits attribute concept hierarchies that represent ISA relationships among domain concepts. It begins with the most generalized summary and proceeds to find more specialized ones by stepwise refinements. This top-down paradigm reveals at least two important advantages compared to the previous bottom-up methods. Firstly, it provides a natural way of reflecting the user's own discovery preference interactively. Secondly, it does not produce too large intermediate result that makes it hard for the bottom-up approach to be applied in practical environment. The proposed procedure can also be easily extended for distributed databases. Information content measure of a database summary is derived in order to identify more informative summaries among the discovered results.

A Column-Aware Index Management Using Flash Memory for Read-Intensive Databases

  • Byun, Si-Woo;Jang, Seok-Woo
    • Journal of Information Processing Systems
    • /
    • v.11 no.3
    • /
    • pp.389-405
    • /
    • 2015
  • Most traditional database systems exploit a record-oriented model where the attributes of a record are placed contiguously in a hard disk to achieve high performance writes. However, for read-mostly data warehouse systems, the column-oriented database has become a proper model because of its superior read performance. Today, flash memory is largely recognized as the preferred storage media for high-speed database systems. In this paper, we introduce a column-oriented database model based on flash memory and then propose a new column-aware flash indexing scheme for the high-speed column-oriented data warehouse systems. Our index management scheme, which uses an enhanced $B^+$-Tree, achieves superior search performance by indexing an embedded segment and packing an unused space in internal and leaf nodes. Based on the performance results of two test databases, we concluded that the column-aware flash index management outperforms the traditional scheme in the respect of the mixed operation throughput and its response time.

Analysis of Impact Between Data Analysis Performance and Database

  • Kyoungju Min;Jeongyun Cho;Manho Jung;Hyangbae Lee
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.3
    • /
    • pp.244-251
    • /
    • 2023
  • Engineering or humanities data are stored in databases and are often used for search services. While the latest deep-learning technologies, such like BART and BERT, are utilized for data analysis, humanities data still rely on traditional databases. Representative analysis methods include n-gram and lexical statistical extraction. However, when using a database, performance limitation is often imposed on the result calculations. This study presents an experimental process using MariaDB on a PC, which is easily accessible in a laboratory, to analyze the impact of the database on data analysis performance. The findings highlight the fact that the database becomes a bottleneck when analyzing large-scale text data, particularly over hundreds of thousands of records. To address this issue, a method was proposed to provide real-time humanities data analysis web services by leveraging the open source database, with a focus on the Seungjeongwon-Ilgy, one of the largest datasets in the humanities fields.

Performance Enhancement Architecture including Location Information Secrecy for HLR System (위치 정보의 보안성이 고려된 가입자 위치등록기 시스템의 새로운 구조)

  • 김자환
    • Proceedings of the Korea Information Assurance Society Conference
    • /
    • 2004.05a
    • /
    • pp.103-108
    • /
    • 2004
  • A Home Location Register(HLR) database system manages each subscriber's location information, which continuously changes in a cellular network. For this purpose, the HLR database system provides table management, index management, and backup management facilities. In this thesis, I propose using a two-level index method for the mobile directory number(MDN) as a suitable method and a chained bucket hashing method for the electronic serial number(ESN). Both the MDN and the ESN are used as keys in the HLR database system. I also propose an efficient backup method that takes into account the characteristics of HLR database transactions. The retrieval speed and the memory usage of the two-level index method are better than those of the T-tree index method. The insertion and deletion overhead of the chained bucket hashing method is less than that of the modified linear hashing method. In the proposed backup method, I use two kinds of dirty flags in order to solvethe performance degradation problem caused by frequent registration-location operations. I also propose using additional attributes in the HLR database scheme for location information secrecy as a suitable security method.

  • PDF

Geological Map Database Construction Using GIS (GIS를 이용한 지질도 데이터베이스 구축에 관한 연구)

  • 이사로;최위찬;민경덕
    • Spatial Information Research
    • /
    • v.7 no.1
    • /
    • pp.147-153
    • /
    • 1999
  • Geological map and data are needed for land use planning, resources development, geological hazard prevention, environment protection and education, Since the nationwide geological database in Korea has not been constructed yet, there are many problems in using the geological map and data. There are many problems such a stratigraphy unestablishment, map conservation and edge matching in geological paper map. Therefore it is difficult to construct the geological map database, but the geological map database must be constructed as soon as possible as one of national thematic map. In this study, geological maps of pilot area such as Ansung geological map on a scale of 1:50,000, Busan on a scale of 1:250,000, Namchang on a scale of 1:25,000 and the whole Korean peninsula on a scale of 1:1,000,000 were designed and constructed to database using Geographic Information System(GIS). In addition the geological map management program was developed by GIS program. The digital geological maps were produced using the constructed geological database. The database could be of access through Internet World Wide Web(WWW) environment and be distributed in Compact Disk(CD).

  • PDF

Resiliency Design of a Loosely-Coupled Database System

  • Park, Jae-Hwa;Kim, Sung-Eon
    • The Journal of Information Systems
    • /
    • v.4
    • /
    • pp.85-104
    • /
    • 1995
  • In a loosely-coupled distributed database server system, a server failure and/or a communication failure can be masked by a resiliency mechanism. Recognizing that a distributed transaction executes at several servers during its lifetime, we propose a resiliency mechanism which allows continuous transaction processing in distributed database server systems in the presence of a server failure. The resiliency mechanism for transaction processing is achieved by keeping redundant information using a primary/backup approach. The purpose of this paper is to analyze the performance improvement opportunities with the resiliency mechanism and to present the design of the proposed system.

  • PDF