Performance of Memory Virtualization Using Global Memory Resource Balancing

2019 ◽  
Vol 9 (1) ◽  
pp. 16-32 ◽  
Author(s):  
Pvss Gangadhar ◽  
Ashok Kumar Hota ◽  
Mandapati Venkateswara Rao ◽  
Vedula Venkateswara Rao

Virtualization has become a universal generalization layer in contemporary data centers. By multiplexing hardware resources into multiple virtual machines and facilitating several operating systems to run on the same physical platform at the same time, it can effectively decrease power consumption and building size or improve security by isolating virtual machines. In a virtualized system, memory resource supervision acts as a decisive task in achieving high resource employment and performance. Insufficient memory allocation to a virtual machine will degrade its performance drastically. On the contrasting, over allocation reasons ravage of memory resources. In the meantime, a virtual machine's memory stipulates may differ drastically. As a consequence, effective memory resource management calls for a dynamic memory balancer, which, preferably, can alter memory allocation in a timely mode for each virtual machine-based on their present memory stipulate and therefore realize the preeminent memory utilization and the best possible overall performance. Migrating operating system instances across discrete physical hosts is a helpful tool for administrators of data centers and clusters: It permits a clean separation among hardware and software, and make easy fault management. In order to approximate the memory, the stipulate of each virtual machine and to adjudicate probable memory resource disagreement, an extensively planned approach is to build an Least Recently Used based miss ratio curve which provides not only the current working set size but also the correlation between performance and the target memory allocation size. In this paper, the authors initially present a low overhead LRU-based memory demand tracking scheme, which includes three orthogonal optimizations: AVL based Least Recently Used association, dynamic hot set sizing. This assessment outcome confirms that, for the complete SPEC CPU 2006 benchmark set, subsequent to pertaining the 3 optimizing techniques, the mean overhead of MRC construction are lowered from 173% to only 2%. Based on current WSS, the authors then predict its trend in the near future and take different tactics for different forecast results. When there is an adequate amount of physical memory on the host, it locally balances its memory resource for the VMs. Once the local memory resource is insufficient and the memory pressure is predicted to sustain for a sufficiently long time, VM live migration, is used to move one or more VMs from the hot host to other host(s). Finally, for transient memory pressure, a remote cache is used to alleviate the temporary performance penalty. These experimental results show that this design achieves 49% center-wide speedup.

2020 ◽  
Vol 2020 ◽  
pp. 1-13
Author(s):  
Xialin Liu ◽  
Junsheng Wu ◽  
Gang Sha ◽  
Shuqin Liu

Cloud data centers consume huge amount of electrical energy bringing about in high operating costs and carbon dioxide emissions. Virtual machine (VM) consolidation utilizes live migration of virtual machines (VMs) to transfer a VM among physical servers in order to improve the utilization of resources and energy efficiency in cloud data centers. Most of the current VM consolidation approaches tend to aggressive-migrate for some types of applications such as large capacity application such as speech recognition, image processing, and decision support systems. These approaches generate a high migration thrashing because VMs are consolidated to servers according to VM’s instant resource usage without considering their overall and long-term utilization. The proposed approach, dynamic consolidation with minimization of migration thrashing (DCMMT) which prioritizes VM with high capacity, significantly reduces migration thrashing and the number of migrations to ensure service-level agreement (SLA) since it keeps VMs likely to suffer from migration thrashing in the same physical servers instead of migrating. We have performed experiments using real workload traces compared to existing aggressive-migration-based solutions; through simulations, we show that our approach improves migration thrashing metric by about 28%, number of migrations metric by about 21%, and SLAV metric by about 19%.


Sensors ◽  
2019 ◽  
Vol 19 (12) ◽  
pp. 2724 ◽  
Author(s):  
Yuan ◽  
Sun

High-energy consumption in data centers has become a critical issue. The dynamic server consolidation has significant effects on saving energy of a data center. An effective way to consolidate virtual machines is to migrate virtual machines in real time so that some light load physical machines can be turned off or switched to low-power mode. The present challenge is to reduce the energy consumption of cloud data centers. In this paper, for the first time, a server consolidation algorithm based on the culture multiple-ant-colony algorithm was proposed for dynamic execution of virtual machine migration, thus reducing the energy consumption of cloud data centers. The server consolidation algorithm based on the culture multiple-ant-colony algorithm (CMACA) finds an approximate optimal solution through a specific target function. The simulation results show that the proposed algorithm not only reduces the energy consumption but also reduces the number of virtual machine migration.


2019 ◽  
Vol 16 (4) ◽  
pp. 627-637
Author(s):  
Sanaz Hosseinzadeh Sabeti ◽  
Maryam Mollabgher

Goal: Load balancing policies often map workloads on virtual machines, and are being sought to achieve their goals by creating an almost equal level of workload on any virtual machine. In this research, a hybrid load balancing algorithm is proposed with the aim of reducing response time and processing time. Design / Methodology / Approach: The proposed algorithm performs load balancing using a table including the status indicators of virtual machines and the task list allocated to each virtual machine. The evaluation results of response time and processing time in data centers from four algorithms, ESCE, Throttled, Round Robin and the proposed algorithm is done. Results: The overall response time and data processing time in the proposed algorithm data center are shorter than other algorithms and improve the response time and data processing time in the data center. The results of the overall response time for all algorithms show that the response time of the proposed algorithm is 12.28%, compared to the Round Robin algorithm, 9.1% compared to the Throttled algorithm, and 4.86% of the ESCE algorithm. Limitations of the investigation: Due to time and technical limitations, load balancing has not been achieved with more goals, such as lowering costs and increasing productivity. Practical implications: The implementation of a hybrid load factor policy can improve the response time and processing time. The use of load balancing will cause the traffic load between virtual machines to be properly distributed and prevent bottlenecks. This will be effective in increasing customer responsiveness. And finally, improving response time increases the satisfaction of cloud users and increases the productivity of computing resources. Originality/Value: This research can be effective in optimizing the existing algorithms and will take a step towards further research in this regard.


2021 ◽  
Vol 39 (1B) ◽  
pp. 203-208
Author(s):  
Haider A. Ghanem ◽  
Rana F. Ghani ◽  
Maha J. Abbas

Data centers are the main nerve of the Internet because of its hosting, storage, cloud computing and other services. All these services require a lot of work and resources, such as energy and cooling. The main problem is how to improve the work of data centers through increased resource utilization by using virtual host simulations and exploiting all server resources. In this paper, we have considered memory resources, where Virtual machines were distributed to hosts after comparing the virtual machines with the host from where the memory and putting the virtual machine on the appropriate host, this will reduce the host machines in the data centers and this will improve the performance of the data centers, in terms of power consumption and the number of servers used and cost.


2013 ◽  
Vol 2013 ◽  
pp. 1-10 ◽  
Author(s):  
Xiaoying Wang ◽  
Xiaojing Liu ◽  
Lihua Fan ◽  
Xuhan Jia

As cloud computing offers services to lots of users worldwide, pervasive applications from customers are hosted by large-scale data centers. Upon such platforms, virtualization technology is employed to multiplex the underlying physical resources. Since the incoming loads of different application vary significantly, it is important and critical to manage the placement and resource allocation schemes of the virtual machines (VMs) in order to guarantee the quality of services. In this paper, we propose a decentralized virtual machine migration approach inside the data centers for cloud computing environments. The system models and power models are defined and described first. Then, we present the key steps of the decentralized mechanism, including the establishment of load vectors, load information collection, VM selection, and destination determination. A two-threshold decentralized migration algorithm is implemented to further save the energy consumption as well as keeping the quality of services. By examining the effect of our approach by performance evaluation experiments, the thresholds and other factors are analyzed and discussed. The results illustrate that the proposed approach can efficiently balance the loads across different physical nodes and also can lead to less power consumption of the entire system holistically.


2020 ◽  
Author(s):  
Swasthi Shetty ◽  
Annappa B

<div> <div> <div> <p>Virtual machine consolidation techniques provide ways to save energy and cost in cloud data centers. However, aggressive packing of virtual machines can cause performance degradation. Therefore, it is essential to strike a trade-off between energy and performance in data centers. Achieving this trade-off has been an active research area in recent years. In this paper, a host underload detection algorithm and a new VM selection and VM placement techniques are proposed to consolidate Virtual machines based on the growth potential of VMs. Growth potential is calculated based on the utilization history of VMs. The interdependence of VM selection and VM placement techniques are also studied in the proposed model. The proposed algorithms are evaluated on real- world PlanetLab workload on Cloudsim. The experimental evaluation shows that our proposed technique reduces Service Level Agreement Violation (SLAV) and energy consumption compared to the existing algorithms. </p> </div> </div> </div>


Cloud computing, with its great potential in low cost and demanding services, is a good computing platform. Modern data centers for cloud computing are facing the difficulty of consistently increasing complexity because of the expanding quantity of clients and their enlarging resource demands. A great deal of efforts are currently focused on giving the cloud framework with autonomic behavior , so it can take decision about virtual machine (VM) management over the datacenter without intervention of human beings. Most of the self-organizing solutions results in eager migration, which attempts to diminish the amount of working servers virtual machines. These self-organizing resolution produce needless migration due to unpredictable workload. So also it consume huge amounts of electrical energy during unnecessary migration process. To overcome this issue, this project develop one novel VM migration scheme called eeadSelfCloud. The proposed schema is used to change the virtual machine in a cloud center that requires a lot of factors, such as basic requirements for resources during virtual machine setup, dynamic resource allocation, top software loading, software execution, and power saving at the Data Center. Data Center Utilization, Average Node Utilization, Request Rejection Ration, Number of Hop Count and Power Consumption are taken as constraint for measuring the proposed approach. The analysis report depicted that the proposed approach performs best than the other existing approaches.


Author(s):  
Arnab Kumar Paul ◽  
Bibhudatta Sahoo

The aim of cloud computing is to enable users to access resources on demand. The number of users is continuously increasing. In order to fulfil their needs, we need more number of physical machines and data centers. The increase in the number of physical machines is directly proportional to the consumption of energy. This gives us one of the major challenges; minimization of energy consumption. One of the most effective ways to minimize the consumption of energy is the optimal virtual machine placement on physical machines. This chapter focuses on finding the solution to the problem of dynamic virtual machine placement for the optimized consumption of energy. An energy consumption model is built which takes into account the states of physical machines and live migration of virtual machines. On top of this, the cloud computing model is built. Unlike centralized approaches towards virtual machine placement which result in many unreachable solutions, a decentralized approach is used in this chapter which provides a list of virtual machine migrations for their optimal placement.


Sign in / Sign up

Export Citation Format

Share Document