Database Replica

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 1656 Experts worldwide ranked by ideXlab platform

Guillaume Pierre - One of the best experts on this subject based on the ideXlab platform.

  • WWW 2007 / Track: Performance and Scalability Session: Scalable Systems for Dynamic Content GlobeTP: Template-Based Database Replication for Scalable Web Applications
    2014
    Co-Authors: Tobias Groothuyse Sivasubramanian, De Boelelaan A, Guillaume Pierre
    Abstract:

    Generic Database Replication algorithms do not scale linearly in throughput as all update, deletion and insertion (UDI) queries must be applied to every Database Replica. The throughput is therefore limited to the point where the number of UDI queries alone is sufficient to overload one server. In such scenarios, partial Replication of a Database can help, as UDI queries are executed only by a subset of all servers. In this paper we propose GlobeTP, a system that employs partial Replication to improve Database throughput. GlobeTP exploits the fact that a Web application’s query workload is composed of a small set of read and write templates. Using knowledge of these templates and their respective execution costs, GlobeTP provides Database table placements that produce significant improvements in Database throughput. We demonstrate the efficiency of this technique using two different industry standard benchmarks. In our experiments, GlobeTP increases the throughput by 57 % to 150% compared to full Replication, while using identical hardware configuration. Furthermore, adding a single query cache improves the throughput by another 30 % to 60%. Categories and Subject Descriptor

  • globetp template based Database Replication for scalable web applications
    2007
    Co-Authors: Tobias Groothuyse, Swaminathan Sivasubramanian, Guillaume Pierre
    Abstract:

    Generic Database Replication algorithms do not scale linearly in throughput as all update, deletion and insertion (UDI) queries must be applied to every Database Replica. The throughput is therefore limited to the point where the number of UDI queries alone is sufficient to overload one server. In such scenarios, partial Replication of a Database can help, as UDI queries are executed only by a subset of all servers. In this paper we propose GlobeTP, a system that employs partial Replication to improve Database throughput. GlobeTP exploits the fact that a Web application's query workload is composed of a small set of read and write templates. Using knowledge of these templates and their respective execution costs, GlobeTP provides Database table placements that produce significant improvements in Database throughput. We demonstrate the efficiency of this technique using two different industry standard benchmarks. In our experiments, GlobeTP increases the throughput by 57% to 150% compared to full Replication, while using identical hardware configuration. Furthermore, adding a single query cache improves the throughput by another 30% to 60%.

Cincilla Pierpaolo - One of the best experts on this subject based on the ideXlab platform.

  • Gargamel: accroître les performances des DBMS en parallélisant les transactions en écriture
    2014
    Co-Authors: Cincilla Pierpaolo
    Abstract:

    Databases often scale poorly in distributed configurations, due to the cost of concurrencycontrol and to resource contention. The alternative of centralizing writes workswell only for read-intensive workloads, whereas weakening transactional properties isproblematic for application developers. Our solution spreads non-conflicting updatetransactions to different Replicas, but still provides strong transactional guarantees. Ineffect, Gargamel partitions the Database dynamically according to the update workload.Each Database Replica runs sequentially, at full bandwidth; mutual synchronisation betweenReplicas remains minimal. Our prototype show that Gargamel improves both responsetime and load by an order of magnitude when contention is high (highly loadedsystem with bounded resources), and that otherwise slow-down is negligible.Les bases de données présentent des problèmes de passage à l’échelle. Ceci estprincipalement dû à la compétition pour les ressources et au coût du contrôle de laconcurrence. Une alternative consiste à centraliser les écritures afin d’éviter les conflits.Cependant, cette solution ne présente des performances satisfaisantes que pour lesapplications effectuant majoritairement des lectures. Une autre solution est d’affaiblirles propriétés transactionnelles mais cela complexifie le travail des développeursd’applications. Notre solution, Gargamel, répartie les transactions effectuant des écrituressur différentes répliques de la base de données tout en gardant de fortes propriétéstransactionnelles. Toutes les répliques de la base de donnée s’exécutent séquentiellement,à plein débit; la synchronisation entre les répliques reste minime. Les évaluationseffectuées avec notre prototype montrent que Gargamel permet d’améliorer le temps deréponse et la charge d’un ordre de grandeur quand la compétition est forte (systèmestrès chargés avec ressources limitées) et que dans les autres cas le ralentissement estnégligeable

  • Gargamel : boosting DBMS performance by parallelising write transactions
    2014
    Co-Authors: Cincilla Pierpaolo
    Abstract:

    Databases often scale poorly in distributed configurations, due to the cost of concurrency control and to resource contention. The alternative of centralizing writes works well only for read-intensive workloads, whereas weakening transactional properties is problematic for application developers. Our solution spreads non-conflicting update transactions to different Replicas, but still provides strong transactional guarantees. In effect, Gargamel partitions the Database dynamically according to the update workload. Each Database Replica runs sequentially, at full bandwidth; mutual synchronisation between Replicas remains minimal. Our prototype show that Gargamel improves both response time and load by an order of magnitude when contention is high (highly loaded system with bounded resources), and that otherwise slow-down is negligible.Les bases de données présentent des problèmes de passage à l’échelle. Ceci est principalement dû à la compétition pour les ressources et au coût du contrôle de la concurrence. Une alternative consiste à centraliser les écritures afin d’éviter les conflits. Cependant, cette solution ne présente des performances satisfaisantes que pour les applications effectuant majoritairement des lectures. Une autre solution est d’affaiblir les propriétés transactionnelles mais cela complexifie le travail des développeurs d’applications. Notre solution, Gargamel, répartie les transactions effectuant des écritures sur différentes répliques de la base de données tout en gardant de fortes propriétés transactionnelles. Toutes les répliques de la base de donnée s’exécutent séquentiellement, à plein débit; la synchronisation entre les répliques reste minime. Les évaluations effectuées avec notre prototype montrent que Gargamel permet d’améliorer le temps de réponse et la charge d’un ordre de grandeur quand la compétition est forte (systèmes très chargés avec ressources limitées) et que dans les autres cas le ralentissement est négligeable

Alok Pareek - One of the best experts on this subject based on the ideXlab platform.

  • bronzegate real time transactional data obfuscation for goldengate
    2010
    Co-Authors: Shenoda Guirguis, Alok Pareek
    Abstract:

    Data privacy laws have appeared recently, such as the HIPAA laws for protecting medical records, and the PCI guidelines for protecting Credit Card information. Data privacy can be defined as maintaining the privacy of Personal Identifiable Information (PII) from unauthorized accessing. PII includes any piece of data that can be used alone, or in conjunction with additional information, to uniquely identify an individual. Examples of such information include national identification numbers, credit card numbers, as well as financial and medical records. Access control methods and data encryption provide a level of data protection from unauthorized access, however, it is not enough; it does not prohibit identity thefts. It was reported that 70% of the data privacy breaches are internal breaches that involve an employee from the enterprise who has access to some training or testing Database Replica, which contains all the PII. In addition to access control, we need techniques to obfuscate (i.e., mask or dim) the datasets used for training, testing and analysis purposes. A good data obfuscation technique would, among other features, preserve the data usability while protecting its privacy. This challenge is further complicated when real time requirements are added. In this paper we present BronzeGate: Obfuscated GoldenGate, the GoldenGate's real-time solution for transactional data privacy while maintaining data usability. BronzeGate utilizes different obfuscation functions for different data types to securely obfuscate the data, on real-time, while maintaining its statistical characteristics.

Tobias Groothuyse - One of the best experts on this subject based on the ideXlab platform.

  • globetp template based Database Replication for scalable web applications
    2007
    Co-Authors: Tobias Groothuyse, Swaminathan Sivasubramanian, Guillaume Pierre
    Abstract:

    Generic Database Replication algorithms do not scale linearly in throughput as all update, deletion and insertion (UDI) queries must be applied to every Database Replica. The throughput is therefore limited to the point where the number of UDI queries alone is sufficient to overload one server. In such scenarios, partial Replication of a Database can help, as UDI queries are executed only by a subset of all servers. In this paper we propose GlobeTP, a system that employs partial Replication to improve Database throughput. GlobeTP exploits the fact that a Web application's query workload is composed of a small set of read and write templates. Using knowledge of these templates and their respective execution costs, GlobeTP provides Database table placements that produce significant improvements in Database throughput. We demonstrate the efficiency of this technique using two different industry standard benchmarks. In our experiments, GlobeTP increases the throughput by 57% to 150% compared to full Replication, while using identical hardware configuration. Furthermore, adding a single query cache improves the throughput by another 30% to 60%.

Tobias Groothuyse Sivasubramanian - One of the best experts on this subject based on the ideXlab platform.

  • WWW 2007 / Track: Performance and Scalability Session: Scalable Systems for Dynamic Content GlobeTP: Template-Based Database Replication for Scalable Web Applications
    2014
    Co-Authors: Tobias Groothuyse Sivasubramanian, De Boelelaan A, Guillaume Pierre
    Abstract:

    Generic Database Replication algorithms do not scale linearly in throughput as all update, deletion and insertion (UDI) queries must be applied to every Database Replica. The throughput is therefore limited to the point where the number of UDI queries alone is sufficient to overload one server. In such scenarios, partial Replication of a Database can help, as UDI queries are executed only by a subset of all servers. In this paper we propose GlobeTP, a system that employs partial Replication to improve Database throughput. GlobeTP exploits the fact that a Web application’s query workload is composed of a small set of read and write templates. Using knowledge of these templates and their respective execution costs, GlobeTP provides Database table placements that produce significant improvements in Database throughput. We demonstrate the efficiency of this technique using two different industry standard benchmarks. In our experiments, GlobeTP increases the throughput by 57 % to 150% compared to full Replication, while using identical hardware configuration. Furthermore, adding a single query cache improves the throughput by another 30 % to 60%. Categories and Subject Descriptor