Database Replication

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 11358 Experts worldwide ranked by ideXlab platform

Bettina Kemme - One of the best experts on this subject based on the ideXlab platform.

  • Database Replication
    2010
    Co-Authors: Bettina Kemme, Ricardo Jimnez Peris, Marta Patio-martnez
    Abstract:

    Database Replication is widely used for fault-tolerance, scalability and performance. The failure of one Database replica does not stop the system from working as available replicas can take over the tasks of the failed replica. Scalability can be achieved by distributing the load across all replicas, and adding new replicas should the load increase. Finally, Database Replication can provide fast local access, even if clients are geographically distributed clients, if data copies are located close to clients. Despite its advantages, Replication is not a straightforward technique to apply, and there are many hurdles to overcome. At the forefront is replica control: assuring that data copies remain consistent when updates occur. There exist many alternatives in regard to where updates can occur and when changes are propagated to data copies, how changes are applied, where the Replication tool is located, etc. A particular challenge is to combine replica control with transaction management as it requires several operations to be treated as a single logical unit, and it provides atomicity, consistency, isolation and durability across the replicated system. The book provides a categorization of replica control mechanisms, presents several replica and concurrency control mechanisms in detail, and discusses many of the issues that arise when such solutions need to be implemented within or on top of relational Database systems. Table of Contents: Overview / 1-Copy-Equivalence and Consistency / Basic Protocols / Replication Architecture / The Scalability of Replication / Eager Replication and 1-Copy-Serializability / 1-Copy-Snapshot Isolation / Lazy Replication / Self-Configuration and Elasticity / Other Aspects of Replication

  • Database Replication: a tale of research across communities
    2010
    Co-Authors: Bettina Kemme, Gustavo Alonso
    Abstract:

    Replication is a key mechanism to achieve scalability and fault-tolerance in Databases. Its importance has recently been further increased because of the role it plays in achieving elasticity at the Database layer. In Database Replication, the biggest challenge lies in the trade-off between performance and consistency. A decade ago, performance could only be achieved through lazy Replication at the expense of transactional guarantees. The strong consistency of eager approaches came with a high cost in terms of reduced performance and limited scalability. Postgres-R combined results from distributed systems and Databases to develop a Replication solution that provided both scalability and strong consistency. The use of group communication primitives with strong ordering and delivery guarantees together with optimized transaction handling (tailored locking, transferring logs instead of re-executing updates, keeping the message overhead per transaction constant) were a drastic departure from the state-of-the-art at the time. Ten years later, these techniques are widely used in a variety of contexts but particularly in cloud computing scenarios. In this paper we review the original motivation for Postgres-R and discuss how the ideas behind the design have evolved over the years.

  • Replication - Database Replication: a tutorial
    Lecture Notes in Computer Science, 2010
    Co-Authors: Bettina Kemme, Marta Patiño-martínez, Ricardo Jiménez-peris, Gustavo Alonso
    Abstract:

    This chapter provides an in-depth introduction to Database Replication, in particular how transactions are executed in a replicated environment. We describe a suite of Replication protocols and illustrate the design alternatives using a two-step approach. We first categorize Replication protocols by only two parameters and present a simple example protocol for each of the resulting categories. Further parameters are then introduced, and we illustrate them by the given Replication protocols and some variants.

  • boosting Database Replication scalability through partial Replication and 1 copy snapshot isolation
    Pacific Rim International Symposium on Dependable Computing, 2007
    Co-Authors: Damian Serrano, Marta Patinomartinez, Ricardo Jimenezperis, Bettina Kemme
    Abstract:

    Databases have become a crucial component in modern information systems. At the same time, they have become the main bottleneck in most systems. Database Replication protocols have been proposed to solve the scalability problem by scaling out in a cluster of sites. Current techniques have attained some degree of scalability, however there are two main limitations to existing approaches. Firstly, most solutions adopt a full Replication model where all sites store a full copy of the Database. The coordination overhead imposed by keeping all replicas consistent allows such approaches to achieve only medium scalability. Secondly, most Replication protocols rely on the traditional consistency criterion, 1-copy-serializability, which limits concurrency, and thus scalability of the system. In this paper, we first analyze analytically the performance gains that can be achieved by various partial Replication configurations, i.e., configurations where not all sites store all data. From there, we derive a partial Replication protocol that provides 1-copy-snapshot isolation as correctness criterion. We have evaluated the protocol with TPC-W and the results show better scalability than full Replication.

  • PRDC - Boosting Database Replication Scalability through Partial Replication and 1-Copy-Snapshot-Isolation
    13th Pacific Rim International Symposium on Dependable Computing (PRDC 2007), 2007
    Co-Authors: Damian Serrano, Ricardo Jiménez-peris, Marta Patiño-martínez, Bettina Kemme
    Abstract:

    Databases have become a crucial component in modern information systems. At the same time, they have become the main bottleneck in most systems. Database Replication protocols have been proposed to solve the scalability problem by scaling out in a cluster of sites. Current techniques have attained some degree of scalability, however there are two main limitations to existing approaches. Firstly, most solutions adopt a full Replication model where all sites store a full copy of the Database. The coordination overhead imposed by keeping all replicas consistent allows such approaches to achieve only medium scalability. Secondly, most Replication protocols rely on the traditional consistency criterion, 1-copy-serializability, which limits concurrency, and thus scalability of the system. In this paper, we first analyze analytically the performance gains that can be achieved by various partial Replication configurations, i.e., configurations where not all sites store all data. From there, we derive a partial Replication protocol that provides 1-copy-snapshot isolation as correctness criterion. We have evaluated the protocol with TPC-W and the results show better scalability than full Replication.

Cipria Tutu - One of the best experts on this subject based on the ideXlab platform.

  • On the Path from Total Order to Database Replication
    2004
    Co-Authors: Yair Amir, Cipria Tutu
    Abstract:

    We introduce ZBCast, a primitive that provides Persistent Global Total Order for messages delivered to a group of participants that can crash and subsequently recover and that can become temporarily partitioned and then remerge due to network conditions. The paper presents in detail and proves the correctness of an efficient algorithm that implements ZBCast on top of existing group communication infrastructure. The algorithm minimizes the amount of required forced disk writes and avoids the need for application level (end-to-end) acknowledgments per message. We also present an extension of the algorithm that allows dynamic addition or removal of participants. We indicate how ZBCast can be employed to build a generic data Replication engine that can be used to provide consistent synchronous Database Replication. We provide experimental results that indicate the efficiency of the approach.

  • on the performance of consistent wide area Database Replication
    2003
    Co-Authors: Yair Amirclaudiu Danilovmichal Miskinamir, Jonathan Stanton, Cipria Tutu
    Abstract:

    In this paper we design a generic, consistent Replication architecture that enables transparent Database Replication and we present the optimizations and tradeoffs of the chosen design. We demonstrate the practicality of our approach by building a prototype that replicates a PostgreSQL Database system. We provide experimental results for consistent wide-area Database Replication. We claim that the use of an optimized synchronization engine is the key to building a practical synchronous Replication system for wide-area network settings.

  • from total order to Database Replication
    International Conference on Distributed Computing Systems, 2002
    Co-Authors: Yair Amir, Cipria Tutu
    Abstract:

    This paper presents in detail an efficient and provably correct algorithm for Database Replication over partitionable networks. Our algorithm avoids the need for end-to-end acknowledgments for each action while supporting network partitions and merges and allowing dynamic instantiation of new replicas. One round of end-to-end acknowledgments is required only upon a membership change event such as a network partition. New actions may be introduced to the system at any point, not only while in a primary component. We show how performance can be further improved for applications that allow relaxation of consistency requirements. We provide experimental results that demonstrate the efficiency of our approach.

  • on the performance of wide area synchronous Database Replication
    2002
    Co-Authors: Yai Ami, Claudiu Danilov, Michal Miskinami, Jonatha Stanto, Cipria Tutu
    Abstract:

    A fundamental challenge in Database Replication is to maintain a low cost of updates while assuring global system consistency. The dicult y of the problem is magnied for wide-area network settings due to the high latency and the increased likelihood of network partitions. As a consequence, most of the research in the area has focused either on improving the performance of local transaction execution or on Replication models with weaker semantics, which rely on application knowledge to resolve potential conicts. In this work we identify the performance bottleneck of the existing synchronous Replication schemes as residing in the update synchronization algorithm. We compare the performance of several such synchronization algorithms and highlight the large performance gap between various methods. We design a generic, synchronous Replication scheme that uses an enhanced synchronization algorithm and demonstrate its practicality by building a prototype that replicates a PostgreSQL Database system. We claim that the use of an optimized synchronization engine is the key to building a practical synhronous Replication system for wide-area network settings.

  • Practical Wide-Area Database Replication 1
    2002
    Co-Authors: Yair Amir, Claudiu Danilov, Cipria Tutu, Jonathan Stanton, Michal Miskin-amir, Spread Concepts Llc, Acacia Ave
    Abstract:

    This paper explores the architecture, implementation and performance of a wide and local area Database Replication system. The architecture provides peer Replication, supporting diverse application semantics, based on a group communication paradigm. Network partitions and merges, computer crashes and recoveries, and message omissions are all handled. Using a generic Replication engine and the Spread group communication toolkit, we provide Replication services for the PostgreSQL Database system. We define three different environments to be used as test-beds: a local area cluster, a wide area network that spans the U.S.A, and an emulated wide area test bed. We conduct an extensive set of experiments on these environments, varying the number of replicas and clients, the mix of updates and queries, and the network latency. Our results show that sophisticated algorithms and careful distributed systems design can make symmetric, synchronous, peer Database Replication a reality for both local and wide area networks.

Francesc D. Muñoz-escoí - One of the best experts on this subject based on the ideXlab platform.

  • ICSOFT (Selected Papers) - A Hybrid Approach for Database Replication: Finding the Optimal Configuration between Update Everywhere and Primary Copy Paradigms
    Communications in Computer and Information Science, 2011
    Co-Authors: M. Liroz-gistau, José Enrique Armendáriz-iñigo, J. R. González De Mendívil, J. R. Juárez-rodríguez, Francesc D. Muñoz-escoí
    Abstract:

    Database Replication has been subject of two different approaches, namely primary copy and update everywhere protocols. The former only allows performing update transactions in the primary replica, while the rest are only used to execute read-only transactions. Update everywhere protocols, on the other hand, allow the system to schedule update transactions in any replica, thus increasing its capacity to deal with update intensive workloads and overcoming failures. However, synchronization costs augment and its throughput may fall below the ones obtained by primary copy approaches. Under these circumstances, we propose a new Database Replication paradigm, halfway between primary copy and update everywhere approaches, which improve system’s performance by adapting its configuration depending on the workload submitted to the system. The core of this approach is a deterministic Replication protocol which propagate changes so that broadcast transactions are never aborted. We also propose a recovery algorithm to ensure fault tolerance.

  • Correctness proof of a Database Replication protocol under the perspective of the I/O automaton model
    Acta Informatica, 2009
    Co-Authors: José Enrique Armendáriz-iñigo, José Ramón González De Mendívil, José Ramón Garitagoitia, Francesc D. Muñoz-escoí
    Abstract:

    Correctness of recent Database Replication protocols has been justified in a rather informal way focusing only in safety properties and without using any rigorous formalism. Since a Database Replication protocol must ensure some degree of replica consistency and that transactions follow a given isolation level, previous proofs only focused in these two issues. This paper proposes a formalization using the I/O automaton model, identifying several components in the distributed system that are involved in the Replication support (Replication protocol, group communication system, Database replicas) and specifying clearly their actions in the global replicated system architecture. Then, a general certification-based Replication protocol guaranteeing the snapshot isolation level is proven correct. To this end, different safety and liveness properties are identified, checked and proved. Our work shows that some details of the Replication protocols that were ignored in previous correctness justifications are indeed needed in order to guarantee our proposed correctness criteria.

  • SRDS - Performance Evaluation of a Metaprotocol for Database Replication Adaptability
    2009 28th IEEE International Symposium on Reliable Distributed Systems, 2009
    Co-Authors: M. I. Ruiz-fuertes, Francesc D. Muñoz-escoí
    Abstract:

    Common solutions to Database Replication use a single Replication protocol. This approach lacks flexibility for changing scenarios or when dealing with heterogeneous client application requirements. Our proposal is a metaprotocol that supports several Replication protocols which may follow different Replication techniques or provide different isolation levels. With our metaprotocol, Replication protocols can either work concurrently with the same data or be sequenced for adapting to dynamic environments. Experimental results demonstrate its low overhead and measure the influence of protocol concurrency on system performance.

  • SAC - SIPRe: a partial Database Replication protocol with SI replicas
    Proceedings of the 2008 ACM symposium on Applied computing - SAC '08, 2008
    Co-Authors: José Enrique Armendáriz-iñigo, A. Mauch-goya, J. R. González De Mendívil, Francesc D. Muñoz-escoí
    Abstract:

    Database Replication has been researched as a solution to overcome the problems of performance and availability of distributed systems. Full Database Replication, based on group communication systems, is an attempt to enhance performance that works well for a reduced number of sites. If application locality is taken into consideration, partial Replication, i.e. not all sites store the full Database, also enhances scalability. On the other hand, it is needed to keep all copies consistent. If each DBMS provides SI, the execution of transactions has to be coordinated so as to obtain Generalized-SI (GSI). In this paper, a partial Replication protocol providing GSI is introduced that gives a consistent view of the Database, providing an adaptive Replication technique and supporting the failure and recovery of replicas.

  • OTM Conferences (1) - Extending Middleware Protocols for Database Replication with Integrity Support
    On the Move to Meaningful Internet Systems: OTM 2008, 2008
    Co-Authors: Francesc D. Muñoz-escoí, José Enrique Armendáriz-iñigo, Hendrik Decker, M. I. Ruiz-fuertes, Jose Ramon Gonzalez De Mendivil
    Abstract:

    Current middleware Database Replication protocols take care of read-write conflict evaluation. If there are no such conflicts, protocols sanction transactions to commit. Other conflicts may arise due to integrity violation. So, if, in addition to the consistency of transactions and replicas, also the consistency of integrity constraints is to be supported, some more care must be taken. Some classes of Replication protocols are able to seamlessly deal with the integrity support of the underlying DBMS, but others are not. In this paper, we analyze the support for integrity that can be provided in various classes of Replication protocols. Also, we propose extensions for those that cannot directly manage certain kinds of constraints that are usually supported in DBMSs.

Nuno Preguica - One of the best experts on this subject based on the ideXlab platform.

  • efficient middleware for byzantine fault tolerant Database Replication
    European Conference on Computer Systems, 2011
    Co-Authors: Rui Garcia, Rodrigo Rodrigues, Nuno Preguica
    Abstract:

    Byzantine fault tolerance (BFT) enhances the reliability and availability of replicated systems subject to software bugs, malicious attacks, or other unexpected events. This paper presents Byzantium, a BFT Database Replication middleware that provides snapshot isolation semantics. It is the first BFT Database system that allows for concurrent transaction execution without relying on a centralized component, which is essential for having both performance and robustness. Byzantium builds on an existing BFT library but extends it with a set of techniques for increasing concurrency in the execution of operations, for optimistically executing operations in a single replica, and for striping and load-balancing read operations across replicas. Experimental results show that our Replication protocols introduce only a modest performance overhead for read-write dominated workloads and perform better than a non-replicated Database system for read-only workloads.

  • EuroSys - Efficient middleware for byzantine fault tolerant Database Replication
    Proceedings of the sixth conference on Computer systems - EuroSys '11, 2011
    Co-Authors: Rui Garcia, Rodrigo Rodrigues, Nuno Preguica
    Abstract:

    Byzantine fault tolerance (BFT) enhances the reliability and availability of replicated systems subject to software bugs, malicious attacks, or other unexpected events. This paper presents Byzantium, a BFT Database Replication middleware that provides snapshot isolation semantics. It is the first BFT Database system that allows for concurrent transaction execution without relying on a centralized component, which is essential for having both performance and robustness. Byzantium builds on an existing BFT library but extends it with a set of techniques for increasing concurrency in the execution of operations, for optimistically executing operations in a single replica, and for striping and load-balancing read operations across replicas. Experimental results show that our Replication protocols introduce only a modest performance overhead for read-write dominated workloads and perform better than a non-replicated Database system for read-only workloads.

  • byzantium byzantine fault tolerant Database Replication providing snapshot isolation
    Hot Topics in System Dependability, 2008
    Co-Authors: Nuno Preguica, Rodrigo Rodrigues, Cristovao Honorato, Joao Lourenco
    Abstract:

    Database systems are a key component behind many of today's computer systems. As a consequence, it is crucial that Database systems provide correct and continuous service despite unpredictable circumstances, such as software bugs or attacks. This paper presents the design of Byzantium, a Byzantine fault-tolerant Database Replication middleware that provides snapshot isolation (SI) semantics. SI is very popular because it allows increased concurrency when compared to serializability, while providing similar behavior for typical workloads. Thus, Byzantium improves on existing proposals by allowing increased concurrency and not relying on any centralized component. Our middleware can be used with off-the-shelf Database systems and it is built on top of an existing BFT library.

  • HotDep - Byzantium: Byzantine-fault-tolerant Database Replication providing snapshot isolation
    2008
    Co-Authors: Nuno Preguica, Rodrigo Rodrigues, Cristovao Honorato, Joao Lourenco
    Abstract:

    Database systems are a key component behind many of today's computer systems. As a consequence, it is crucial that Database systems provide correct and continuous service despite unpredictable circumstances, such as software bugs or attacks. This paper presents the design of Byzantium, a Byzantine fault-tolerant Database Replication middleware that provides snapshot isolation (SI) semantics. SI is very popular because it allows increased concurrency when compared to serializability, while providing similar behavior for typical workloads. Thus, Byzantium improves on existing proposals by allowing increased concurrency and not relying on any centralized component. Our middleware can be used with off-the-shelf Database systems and it is built on top of an existing BFT library.

Javam C Machado - One of the best experts on this subject based on the ideXlab platform.

  • towards elastic multi tenant Database Replication with quality of service
    Utility and Cloud Computing, 2012
    Co-Authors: Flavio R C Sousa, Javam C Machado
    Abstract:

    Database systems serving cloud platforms must handle a large number of applications or tenants. Multi-tenancy Database has been prevalent for hosting multiple tenants within a single DBMS while enabling effective resource sharing. Providing such performance goals is challenging for providers as they must balance the performance that they can deliver to tenants and the operating costs. In such Database systems, a key functionality for service providers is Database Replication, which is useful for availability, performance, elasticity, and quality of service. This paper presents RepliC, an approach to Database Replication in the cloud with quality of service, elasticity, and support tomulti-tenancy. In order to evaluate RepliC, some experiments that measure the quality of service and elasticity are presented. Our experiment results confirm that RepliC ensures the quality of service with small SLA violations, while using resources efficiently.

  • UCC - Towards Elastic Multi-Tenant Database Replication with Quality of Service
    2012 IEEE Fifth International Conference on Utility and Cloud Computing, 2012
    Co-Authors: Flavio R C Sousa, Javam C Machado
    Abstract:

    Database systems serving cloud platforms must handle a large number of applications or tenants. Multi-tenancy Database has been prevalent for hosting multiple tenants within a single DBMS while enabling effective resource sharing. Providing such performance goals is challenging for providers as they must balance the performance that they can deliver to tenants and the operating costs. In such Database systems, a key functionality for service providers is Database Replication, which is useful for availability, performance, elasticity, and quality of service. This paper presents RepliC, an approach to Database Replication in the cloud with quality of service, elasticity, and support tomulti-tenancy. In order to evaluate RepliC, some experiments that measure the quality of service and elasticity are presented. Our experiment results confirm that RepliC ensures the quality of service with small SLA violations, while using resources efficiently.