Working Set Size

14,000,000 Leading Edge Experts on the ideXlab platform

Scan Science and Technology

Contact Leading Edge Experts & Companies

Scan Science and Technology

Contact Leading Edge Experts & Companies

The Experts below are selected from a list of 17673 Experts worldwide ranked by ideXlab platform

Ahmed F Amin - One of the best experts on this subject based on the ideXlab platform.

  • more accurate estimation of Working Set Size in virtual machines
    IEEE Access, 2019
    Co-Authors: Ahmed A Harby, Sherif F Fahmy, Ahmed F Amin
    Abstract:

    Accurate Working Set Size estimation is important to increase the consolidation ratio of data centers and to improve the efficiency of live migration. Thus, it is important to come up with a technique that provides an accurate estimation of the Working Set Size of virtual machines that can respond to changes in memory usage in real-time. In this paper, we study the problem of Working Set Size estimation in virtual machines and come up with a method that allows us to better estimate the Working Set Size of virtual machines in Linux. Toward that end, we design a finite state machine that can be used to accurately estimate the Working Set Size and that is responsive to changes in workload. We then implement the algorithm on Linux using QEMU-KVM as our hypervisor. The system is tested using the sysbench benchmark for memory, CPU, and database workloads. The results indicate that our algorithm provides better results in terms of average Working Set Size estimations and is competitive with existing techniques in terms of page faults.

Ahmed A Harby - One of the best experts on this subject based on the ideXlab platform.

  • more accurate estimation of Working Set Size in virtual machines
    IEEE Access, 2019
    Co-Authors: Ahmed A Harby, Sherif F Fahmy, Ahmed F Amin
    Abstract:

    Accurate Working Set Size estimation is important to increase the consolidation ratio of data centers and to improve the efficiency of live migration. Thus, it is important to come up with a technique that provides an accurate estimation of the Working Set Size of virtual machines that can respond to changes in memory usage in real-time. In this paper, we study the problem of Working Set Size estimation in virtual machines and come up with a method that allows us to better estimate the Working Set Size of virtual machines in Linux. Toward that end, we design a finite state machine that can be used to accurately estimate the Working Set Size and that is responsive to changes in workload. We then implement the algorithm on Linux using QEMU-KVM as our hypervisor. The system is tested using the sysbench benchmark for memory, CPU, and database workloads. The results indicate that our algorithm provides better results in terms of average Working Set Size estimations and is competitive with existing techniques in terms of page faults.

Nitu Vlad-tiberiu - One of the best experts on this subject based on the ideXlab platform.

  • Improving energy efficiency of virtualized datacenters
    2018
    Co-Authors: Nitu Vlad-tiberiu
    Abstract:

    Nowadays, many organizations choose to increasingly implement the cloud computing approach. More specifically, as customers, these organizations are outsourcing the management of their physical infrastructure to data centers (or cloud computing platforms). Energy consumption is a primary concern for datacenter (DC) management. Its cost represents about 80% of the total cost of ownership and it is estimated that in 2020, the US DCs alone will spend about $13 billion on energy bills. Generally, the datacenter servers are manufactured in such a way that they achieve high energy efficiency at high utilizations. Thereby for a low cost per computation all datacenter servers should push the utilization as high as possible. In order to fight the historically low utilization, cloud computing adopted server virtualization. The latter allows a physical server to execute multiple virtual servers (called virtual machines) in an isolated way. With virtualization, the cloud provider can pack (consolidate) the entire Set of virtual machines (VMs) on a small Set of physical servers and thereby, reduce the number of active servers. Even so, the datacenter servers rarely reach utilizations higher than 50% which means that they operate with Sets of longterm unused resources (called 'holes'). My first contribution is a cloud management system that dynamically splits/fusions VMs such that they can better fill the holes. This solution is effective only for elastic applications, i.e. applications that can be executed and reconfigured over an arbitrary number of VMs. However the datacenter resource fragmentation stems from a more fundamental problem. Over time, cloud applications demand more and more memory but the physical servers provide more an more CPU. In nowadays datacenters, the two resources are strongly coupled since they are bounded to a physical sever. My second contribution is a practical way to decouple the CPU-memory tuple that can simply be applied to a commodity server. Thereby, the two resources can vary independently, depending on their demand. My third and my forth contribution show a practical system which exploit the second contribution. The underutilization observed on physical servers is also true for virtual machines. It has been shown that VMs consume only a small fraction of the allocated resources because the cloud customers are not able to correctly estimate the resource amount necessary for their applications. My third contribution is a system that estimates the memory consumption (i.e. the Working Set Size) of a VM, with low overhead and high accuracy. Thereby, we can now consolidate the VMs based on their Working Set Size (not the booked memory). However, the drawback of this approach is the risk of memory starvation. If one or multiple VMs have an sharp increase in memory demand, the physical server may run out of memory. This event is undesirable because the cloud platform is unable to provide the client with the booked memory. My fourth contribution is a system that allows a VM to use remote memory provided by a different rack server. Thereby, in the case of a peak memory demand, my system allows the VM to allocate memory on a remote physical server

  • Améliorer l'efficacité énergétique des datacenters virtualisés
    2018
    Co-Authors: Nitu Vlad-tiberiu
    Abstract:

    De nos jours, de nombreuses entreprises choisissent de plus en plus d'adopter le cloud computing. Plus précisément, en tant que clients, elles externalisent la gestion de leur infrastructure physique vers des centres de données (ou plateformes de cloud computing). La consommation d'énergie est une préoccupation majeure pour la gestion des centres de données (datacenter, DC). Son impact financier représente environ 80% du coût total de possession et l'on estime qu'en 2020, les DCs américains dépenseront à eux seuls environ 13 milliards de dollars en factures énergétiques. Généralement, les serveurs de centres de données sont conçus de manière à atteindre une grande efficacité énergétique pour des utilisations élevées. Pour diminuer le coût de calcul, les serveurs de centre de données devraient maximiser leur utilisation. Afin de lutter contre l'utilisation historiquement faible des serveurs, le cloud computing a adopté la virtualisation des serveurs. Cette dernière permet à un serveur physique d'exécuter plusieurs serveurs virtuels (appelés machines virtuelles) de manière isolée. Avec la virtualisation, le fournisseur de cloud peut regrouper (consolider) l'ensemble des machines virtuelles (VM) sur un ensemble réduit de serveurs physiques et ainsi réduire le nombre de serveurs actifs. Même ainsi, les serveurs de centres de données atteignent rarement des utilisations supérieures à 50%, ce qui signifie qu'ils fonctionnent avec des ensembles de ressources majoritairement inutilisées (appelés «trous»). Ma première contribution est un système de gestion de cloud qui divise ou fusionne dynamiquement les machines virtuelles de sorte à ce qu'elles puissent mieux remplir les trous. Cette solution n'est efficace que pour des applications élastiques, c'est-à-dire des applications qui peuvent être exécutées et reconfigurées sur un nombre arbitraire de machines virtuelles. Cependant, la fragmentation des ressources provient d'un problème plus fondamental. On observe que les applications cloud demandent de plus en plus de mémoire, tandis que les serveurs physiques fournissent plus de CPU. Dans les DC actuels, les deux ressources sont fortement couplées puisqu'elles sont liées à un serveur physique. Ma deuxième contribution est un moyen pratique de découpler la paire CPU-mémoire, qui peut être simplement appliquée à n'importe quel serveur. Ainsi, les deux ressources peuvent varier indépendamment, en fonction de leur demande. Ma troisième et ma quatrième contribution montrent un système pratique qui exploite la deuxième contribution. La sous-utilisation observée sur les serveurs physiques existe également pour les machines virtuelles. Il a été démontré que les machines virtuelles ne consomment qu'une petite fraction des ressources allouées car les clients du cloud ne sont pas en mesure d'estimer correctement la quantité de ressources nécessaire à leurs applications. Ma troisième contribution est un système qui estime la consommation de mémoire (c'est-à-dire la taille du Working Set) d'une machine virtuelle, avec un surcoût faible et une grande précision. Ainsi, nous pouvons maintenant consolider les machines virtuelles en fonction de la taille de leur Working Set (plutôt que leur mémoire réservée). Cependant, l'inconvénient de cette approche est le risque de famine de mémoire. Si une ou plusieurs machines virtuelles ont une forte augmentation de la demande en mémoire, le serveur physique peut manquer de mémoire. Cette situation n'est pas souhaitable, car la plate-forme cloud est incapable de fournir au client la mémoire qu'il a payée. Finalement, ma quatrième contribution est un système qui permet à une machine virtuelle d'utiliser la mémoire distante fournie par un autre serveur du rack. Ainsi, dans le cas d'un pic de la demande en mémoire, mon système permet à la VM d'allouer de la mémoire sur un serveur physique distant.Nowadays, many organizations choose to increasingly implement the cloud computing approach. More specifically, as customers, these organizations are outsourcing the management of their physical infrastructure to data centers (or cloud computing platforms). Energy consumption is a primary concern for datacenter (DC) management. Its cost represents about 80% of the total cost of ownership and it is estimated that in 2020, the US DCs alone will spend about 13 billion dollars on energy bills. Generally, the datacenter servers are manufactured in such a way that they achieve high energy efficiency at high utilizations. Thereby for a low cost per computation all datacenter servers should push the utilization as high as possible. In order to fight the historically low utilization, cloud computing adopted server virtualization. The latter allows a physical server to execute multiple virtual servers (called virtual machines) in an isolated way. With virtualization, the cloud provider can pack (consolidate) the entire Set of virtual machines (VMs) on a small Set of physical servers and thereby, reduce the number of active servers. Even so, the datacenter servers rarely reach utilizations higher than 50% which means that they operate with Sets of longterm unused resources (called 'holes'). My first contribution is a cloud management system that dynamically splits/fusions VMs such that they can better fill the holes. This solution is effective only for elastic applications, i.e. applications that can be executed and reconfigured over an arbitrary number of VMs. However the datacenter resource fragmentation stems from a more fundamental problem. Over time, cloud applications demand more and more memory but the physical servers provide more an more CPU. In nowadays datacenters, the two resources are strongly coupled since they are bounded to a physical sever. My second contribution is a practical way to decouple the CPU-memory tuple that can simply be applied to a commodity server. Thereby, the two resources can vary independently, depending on their demand. My third and my forth contribution show a practical system which exploit the second contribution. The underutilization observed on physical servers is also true for virtual machines. It has been shown that VMs consume only a small fraction of the allocated resources because the cloud customers are not able to correctly estimate the resource amount necessary for their applications. My third contribution is a system that estimates the memory consumption (i.e. the Working Set Size) of a VM, with low overhead and high accuracy. Thereby, we can now consolidate the VMs based on their Working Set Size (not the booked memory). However, the drawback of this approach is the risk of memory starvation. If one or multiple VMs have an sharp increase in memory demand, the physical server may run out of memory. This event is undesirable because the cloud platform is unable to provide the client with the booked memory. My fourth contribution is a system that allows a VM to use remote memory provided by a different rack server. Thereby, in the case of a peak memory demand, my system allows the VM to allocate memory on a remote physical server

  • Working Set Size estimation techniques in virtualized environments: One Size does not fit all
    'Association for Computing Machinery (ACM)', 2018
    Co-Authors: Nitu Vlad-tiberiu, Kocharyan Aram, Hannas Yaya Issa, Tchana Alain-bouzaïde, Hagimont Daniel, Astsatryan Hrachya
    Abstract:

    Energy consumption is a primary concern for datacenters? management. Numerous datacenters are relying on virtualization, as it provides flexible resource management means such as virtual machine (VM) checkpoint/restart, migration and consolidation. However, one of the main hindrances to server consolidation is physical memory. In nowadays cloud, memory is generally statically allocated to VMs and wasted if not used. Techniques (such as ballooning) were introduced for dynamically reclaiming memory from VMs, such that only the needed memory is provisioned to each VM. However, the challenge is to precisely monitor the needed memory, i.e., the Working Set of each VM. In this paper, we thoroughly review the main techniques that were proposed for monitoring the Working Set of VMs. Additionally, we have implemented the main techniques in the Xen hypervisor and we have defined different metrics in order to evaluate their efficiency. Based on the evaluation results, we propose Badis, a system which combines several of the existing solutions, using the right solution at the right time. We also propose a consolidation extension which leverages Badis in order to pack the VMs based on the Working Set Size and not the booked memory. The implementation of all techniques, our proposed system, and the benchmarks we have used are publicly available in order to support further research in this domain

Anna Melekhova - One of the best experts on this subject based on the ideXlab platform.

  • machine learning in virtualization estimate a virtual machine s Working Set Size
    International Conference on Cloud Computing, 2013
    Co-Authors: Anna Melekhova
    Abstract:

    Achieving high density of virtual machines on a node while maintaining their performance strongly depends on the correct calculation of a virtual machine's Working Set. Different strategies are applied to solve the problem. Some researchers interpret a virtual machine as an unpredictable memory consumer, while others try to introspect a guest OS's knowledge of memory pressure. This paper introduces a new approach to calculation of the Working Set Size - regression analysis. The technique estimates the memory consumption using a Set of virtualization events. In this investigation, we discuss a correlation between the Working Set Size and virtualization events, demonstrate the applicability of the approach and state its limitations. The argued choice of mathematical instrumentation is given. The collecting of control and learning samples is described in details. The results of final evaluation demonstrate significant resource gain.

Yingwei Luo - One of the best experts on this subject based on the ideXlab platform.

  • low cost Working Set Size tracking
    USENIX Annual Technical Conference, 2011
    Co-Authors: Weiming Zhao, Xinxin Jin, Zhenlin Wang, Xiaolin Wang, Yingwei Luo
    Abstract:

    Efficient memory resource management requires knowledge of the memory demands of applications or systems at runtime. A widely proposed approach is to construct an LRU-based miss ratio curve (MRC), which provides not only the current Working Set Size (WSS) but also the relationship between performance and target memory allocation Size. Unfortunately, the cost of LRUMRC monitoring is nontrivial. Although optimized with AVL-tree based LRU structure and dynamic hot Set sizing, the overhead is still as high as 16% on average. Based on a key insight that for most programs the WSSs are stable most of the time, we design an intermittent tracking scheme, which can temporarily turn off memory tracking when memory demands are predicted to be stable. With the assistance of hardware performance counters, memory tracking can be turned on again if a significant change in memory demands is expected. Experimental results show that, by using this intermittent tracking design, memory tracking can be turned off for 82% of the execution time while the accuracy loss is no more than 4%. More importantly, this design is orthogonal to existing optimizing techniques, such as AVL-tree based LRU structure and dynamic hot Set sizing. By combining the three approaches, the mean overhead is lowered to only 2%. We show that when applied to memory balancing for virtual machines, our scheme brings a speedup of 1.85.