Load Balancing in Very Large Data-Centers: The FISSION Approach

Author(s):  
Tarun Raj ◽  
Tamal Das ◽  
Saurabh Hote ◽  
Ashwin Gumaste
2014 ◽  
Vol 25 (10) ◽  
pp. 2659-2669 ◽  
Author(s):  
Huajie Shao ◽  
Lei Rao ◽  
Zhi Wang ◽  
Xue Liu ◽  
Zhibo Wang ◽  
...  

2021 ◽  
Vol 11 (3) ◽  
pp. 34-48
Author(s):  
J. K. Jeevitha ◽  
Athisha G.

To scale back the energy consumption, this paper proposed three algorithms: The first one is identifying the load balancing factors and redistribute the load. The second one is finding out the most suitable server to assigning the task to the server, achieved by most efficient first fit algorithm (MEFFA), and the third algorithm is processing the task in the server in an efficient way by energy efficient virtual round robin (EEVRR) scheduling algorithm with FAT tree topology architecture. This EEVRR algorithm improves the quality of service via sending the task scheduling performance and cutting the delay in cloud data centers. It increases the energy efficiency by achieving the quality of service (QOS).


2021 ◽  
Vol 11 (3) ◽  
pp. 72-91
Author(s):  
Priyanka H. ◽  
Mary Cherian

Cloud computing has become more prominent, and it is used in large data centers. Distribution of well-organized resources (bandwidth, CPU, and memory) is the major problem in the data centers. The genetically enhanced shuffling frog leaping algorithm (GESFLA) framework is proposed to select the optimal virtual machines to schedule the tasks and allocate them in physical machines (PMs). The proposed GESFLA-based resource allocation technique is useful in minimizing the wastage of resource usage and also minimizes the power consumption of the data center. The proposed GESFL algorithm is compared with task-based particle swarm optimization (TBPSO) for efficiency. The experimental results show the excellence of GESFLA over TBPSO in terms of resource usage ratio, migration time, and total execution time. The proposed GESFLA framework reduces the energy consumption of data center up to 79%, migration time by 67%, and CPU utilization is improved by 9% for Planet Lab workload traces. For the random workload, the execution time is minimized by 71%, transfer time is reduced up to 99%, and the CPU consumption is improved by 17% when compared to TBPSO.


Author(s):  
Xiaojing Hou ◽  
Guozeng Zhao

With the wide application of the cloud computing, the contradiction between high energy cost and low efficiency becomes increasingly prominent. In this article, to solve the problem of energy consumption, a resource scheduling and load balancing fusion algorithm with deep learning strategy is presented. Compared with the corresponding evolutionary algorithms, the proposed algorithm can enhance the diversity of the population, avoid the prematurity to some extent, and have a faster convergence speed. The experimental results show that the proposed algorithm has the most optimal ability of reducing energy consumption of data centers.


2017 ◽  
Vol 83 ◽  
pp. 155-168 ◽  
Author(s):  
Adel Nadjaran Toosi ◽  
Chenhao Qu ◽  
Marcos Dias de Assunção ◽  
Rajkumar Buyya

2018 ◽  
Vol 8 (4) ◽  
pp. 118-133 ◽  
Author(s):  
Fahim Youssef ◽  
Ben Lahmar El Habib ◽  
Rahhali Hamza ◽  
Labriji El Houssine ◽  
Eddaoui Ahmed ◽  
...  

Cloud users can have access to the service based on “pay as you go.” The daily increase of cloud users may decrease the performance, the availability and the profitability of the material and software resources used in cloud service. These challenges were solved by several load balancing algorithms between the virtual machines of the data centers. In order to determine a new load balancing improvement; this article's discussions will be divided into two research axes. The first, the pre-classification of tasks depending on whether their characteristics are accomplished or not (Notion of Levels). This new technique relies on the modeling of tasks classification based on an ascending order using techniques that calculate the worst-case execution time (WCET). The second, the authors choose distributed datacenters between quasi-similar virtual machines and the modeling of relationship between virtual machines using the pre-scheduling levels is included in the data center in terms of standard mathematical functions that controls this relationship. The key point of the improvement, is considering the current load of the virtual machine of a data center and the pre-estimation of the execution time of a task before any allocation. This contribution allows cloud service providers to improve the performance, availability and maximize the use of virtual machines workload in their data centers.


Sign in / Sign up

Export Citation Format

Share Document