• Title/Summary/Keyword: Data replication management

Search Result 60, Processing Time 0.034 seconds

A COMPARATIVE STUDY ON BLOCKCHAIN DATA MANAGEMENT SYSTEMS: BIGCHAINDB VS FALCONDB

  • Abrar Alotaibi;Sarah Alissa;Salahadin Mohammed
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.5
    • /
    • pp.128-134
    • /
    • 2023
  • The widespread usage of blockchain technology in cryptocurrencies has led to the adoption of the blockchain concept in data storage management systems for secure and effective data storage and management. Several innovative studies have proposed solutions that integrate blockchain with distributed databases. In this article, we review current blockchain databases, then focus on two well-known blockchain databases-BigchainDB and FalconDB-to illustrate their architecture and design aspects in more detail. BigchainDB is a distributed database that integrates blockchain properties to enhance immutability and decentralization as well as a high transaction rate, low latency, and accurate queries. Its architecture consists of three layers: the transaction layer, consensus layer, and data model layer. FalconDB, on the other hand, is a shared database that allows multiple clients to collaborate on the database securely and efficiently, even if they have limited resources. It has two layers: the authentication layer and the consensus layer, which are used with client requests and results. Finally, a comparison is made between the two blockchain databases, revealing that they share some characteristics such as immutability, low latency, permission, horizontal scalability, decentralization, and the same consensus protocol. However, they vary in terms of database type, concurrency mechanism, replication model, cost, and the usage of smart contracts.

Finding significant genes using factor analysis (요인 분석을 이용한 유의한 유전자 추출)

  • Lee, Jeong-Wha;Lee, Hye-Seon;Park, Hae-Sang;Jun, Chi-Hyuck
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2006.11a
    • /
    • pp.427-430
    • /
    • 2006
  • Clustering for gene expression data without filtering out noise genes may be distorted or derived inappropriate inference. Identifying significant genes and deleting noise before major analysis is necessary fur meaningful discovery from genes expression pattern. We proposed a new method of finding significant genes using factor analysis which is done on transposed data matrix. We construct significance score that is sum of factor loadings for declared significant number of factor, and set threshold through replication. Our proposed method works well for simulated time-course data for finding significant genes even though variance level gets larger.

  • PDF

Distributed database replicator without locking base relations

  • Lee, Wookey;Kang, Sukho;Park, Jooseok
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 1996.10a
    • /
    • pp.93-95
    • /
    • 1996
  • A replication server is considered to be one of the most effective tools to cope with the problems that may be caused by the complex data replications in distributed database systems. In the distributed environment, locking a table is inevitable and it is the main reason to coerce the system practically. This paper presents an Asynchronous Replicator Scheme (ARS) that basically utilizes the system log as files named differential files to refresh the distributed data files with complicated queries, and that it prevents (normally, huge) base tables from being locked. We take join operations as the complicated queries, not only because the join operation covers almost all the operations, but also because it is one of the most time-consuming and data intensive operations in query processings.

  • PDF

Relocation Strategy for an Efficient Management of Replicated Data on Mobile Computing (모바일 컴퓨팅상에서 중복데이터의 효율적 관리를 위한 재배치 전략)

  • Lee, Byung-Kwan;Oh, Am-Suk;Jeong, Eun-Hee
    • Journal of Korea Multimedia Society
    • /
    • v.7 no.5
    • /
    • pp.689-697
    • /
    • 2004
  • Rapid advances in mobile computing and the availability of wireless communications will soon provide mobile users with the ability to access data regardless of the location of the user or of the data. SRA(Static Replica Allocation) that is traditional scheme has been used for the replication method on the server. This replicates the data on the replica server after a moving host has been transferred to the cell. This strategy is simple and can easily relocate data. However, if a mobile user does not exist in the cell, the replicated data can be deleted in order to maintain data consistency. In addition to, if the mobile host leaves from replicated cell, it is difficult to access data in terms of replication route. Therefore, this paper proposes a new method of relocation based on data consistency strategy called USRAC(User Selection Replica Allocation based on Consistency) and also analyzes access cost according to the moving rate of mobile users, according to the access rate of mobile hosts, and according to the number of cells of mobile users and mobile hosts.

  • PDF

A Study on the Internet Based Monitoring System (인터넷 기반 계측관리시스템에 관한 연구)

  • 한병원;허인욱;조충붕
    • Proceedings of the Korean Geotechical Society Conference
    • /
    • 2002.03a
    • /
    • pp.635-642
    • /
    • 2002
  • In this study, a new internet-based methodology to manage the data acquired during the construction and maintenance of structure is study A new internet-based monitoring system to obtain and to manage the measured information regardless of physical location of each user was developed. This method provided an enhanced integration methodology for construction management by implementing common internet or intranet platform

  • PDF

Update Propagation of Replicated Data in a Peer-to-Peer Environment (Peer-to-Peer 환경에서 중복된 데이터의 갱신 전파 기법)

  • Choi Min-Young;Cho Haeng-Rae
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.4B
    • /
    • pp.311-322
    • /
    • 2006
  • Peer-to-peer (P2P) systems have become a popular medium through which to share huge amounts of data. On the basis of network topology, P2P systems are divided into three types: centralized, structured distribution, unstructured distribution. Unstructured P2P systems such as Gnutella are novel in the sense that they are extensible and reliable. However, as the number of nodes increases, unstructured P2P systems would suffer from the high complexity of search operations that have to scan the network to find the required data items. Efficient replication of data items can reduce the complexity, but it introduces another problem of maintaining consistency among replicated data items when each data item could be updated. In this paper, we propose a new update propagation algorithm that propagates an updated data item to all of its replica. The proposed algorithm can reduce the message transfer overhead by adopting the notion of timestamp and hybrid push/pull messaging.

The Consistency Management Using Trees of Replicated Data Items in Partially Replicated Database (부분 중복 데이터베이스에서 중복 데이터의 트리를 이용한 일관성 유지)

  • Bae, Mi-Sook;Hwang, Bu-Hyun
    • The KIPS Transactions:PartD
    • /
    • v.10D no.4
    • /
    • pp.647-654
    • /
    • 2003
  • The replication of data is used to increase its availability and to improve the performance of a system. The distributed database system has to maintain both the database consistency and the replica consistency. This paper proposes an algorithm which resolves the conflict of the operations by using the mechanism based on the structure that the replicas of each data item are hierarchically organized. Each update is propagated along the tree based on the fact that the root of each data item is the primary replica in partially replicated databases. The use of a hierarchy of data may eliminate useless propagation since the propagation can be done only to sites having the replicas. In consequence, the propagation delay of updates may be reduced. By using the timestamp and a compensating transaction, our algorithm resolves the non-serializability problem caused by the conflict of operations that can happen on the way of the update propagation due to the lazy propagation. This resolution also guarantees the data consistency.

A Classification Model and Case Analyses of Backup Centers for Business Continuity Management (업무지속성관리를 위한 백업센터 유형분류모델과 사례분석)

  • 주우철;최흥식
    • Journal of Information Technology Applications and Management
    • /
    • v.10 no.1
    • /
    • pp.71-89
    • /
    • 2003
  • This paper firstly introduces a classification model for designing the architecture of backup centers in terms of business continuity management. We propose that the replication solution, the level of disaster recovery, and the outsourcing style as three important factors that help explain the architecture of backup centers. We provide the details of each factor of the model and apply them to the real world cases of backup centers in Korea. We conclude that the model is well fitted to many cases of currently available backup centers and can provide backup center architects with a well classified architecture model with specific guidelines.

  • PDF

HyperDB - A High Performance Data Analysis System Based on Grid Computing Technology

  • Kim, Tae-Kyung;Na, Jong-Hwa;Chon, Wan-Sup
    • Journal of the Korean Data and Information Science Society
    • /
    • v.18 no.1
    • /
    • pp.161-174
    • /
    • 2007
  • In this paper, we propose a high performance database cluster system called HyperDB to process OLAP queries efficiently. HyperDB is a virtual database system running on top of internet-connected PCs; the PCs are used for their own purpose at ordinary times, but they are able to participate in the database cluster system at non-office hours. We propose fully logical replication technique and optimal parallel intra-query routing technique for extensibility and performance. Experiment for TPC-R benchmark shows significant performance upgrade compared with conventional approaches.

  • PDF

Methods to Enhance Service Scalability Using Service Replication and Migration (서비스 복제 및 이주를 이용한 서비스 확장성 향상 기법)

  • Kim, Ji-Won;Lee, Jae-Yoo;Kim, Soo-Dong
    • Journal of KIISE:Software and Applications
    • /
    • v.37 no.7
    • /
    • pp.503-517
    • /
    • 2010
  • Service-oriented computing, the effective paradigm for developing service applications by using reusable services, becomes popular. In service-oriented computing, service consumer has no responsibility for managing services, just invokes services what service providers are producing. On the other hand, service providers should manage any resources and data for service consumers can use the service anytime and anywhere. However, it is hard service providers manage the quality of the services because an unspecified number of service consumers. Therefore, service scalability for providing services with higher quality of services specified in a service level agreement becomes a potential problem in service-oriented computing. There have been many researches for scalability in network, database, and distributed computing area. But a research about a definition of service scalability and metrics of measuring service scalability is still not mature in service engineering area. In this paper, we construct a service network which connects multiple service nodes, and integrate all the resources to manage it. And we also present a service scalability framework for managing service scalability by using a mechanism of service migration or replication. In section 3, we, firstly, present the structure of the scalability management framework and basic functionalities. In section 4, we propose scalability enhancement mechanism which is needed to release functionality of the framework. In section 5, we design and implement the framework by using proposed mechanism. In section 6, we demonstrate the result of our case study which dynamically manages services in multi-nodes environment by applying our framework. Through the case study, we show the applicability of our scalability management framework and mechanism.