Realization and Application of Large-scale Fast Optical Circuit Switch for Data Center Networking

Author(s):  
Ken-ichi Sato
2013 ◽  
Vol 25 (16) ◽  
pp. 1589-1592 ◽  
Author(s):  
Nathan Farrington ◽  
Alex Forencich ◽  
George Porter ◽  
P.-C. Sun ◽  
Joseph E. Ford ◽  
...  

Author(s):  
Jiawei Huang ◽  
Shiqi Wang ◽  
Shuping Li ◽  
Shaojun Zou ◽  
Jinbin Hu ◽  
...  

AbstractModern data center networks typically adopt multi-rooted tree topologies such leaf-spine and fat-tree to provide high bisection bandwidth. Load balancing is critical to achieve low latency and high throughput. Although the per-packet schemes such as Random Packet Spraying (RPS) can achieve high network utilization and near-optimal tail latency in symmetric topologies, they are prone to cause significant packet reordering and degrade the network performance. Moreover, some coding-based schemes are proposed to alleviate the problem of packet reordering and loss. Unfortunately, these schemes ignore the traffic characteristics of data center network and cannot achieve good network performance. In this paper, we propose a Heterogeneous Traffic-aware Partition Coding named HTPC to eliminate the impact of packet reordering and improve the performance of short and long flows. HTPC smoothly adjusts the number of redundant packets based on the multi-path congestion information and the traffic characteristics so that the tailing probability of short flows and the timeout probability of long flows can be reduced. Through a series of large-scale NS2 simulations, we demonstrate that HTPC reduces average flow completion time by up to 60% compared with the state-of-the-art mechanisms.


Author(s):  
V. Deeban Chakravarthy ◽  
B. Amutha

Due to the increase in the number of users on the internet and the number of applications that is available in the cloud makes Data Center Networking (DCN) has the backbone for computing. These data centre requires high operational cost and also experience the link failures and congestions often. Hence the solution is to use Software Defined Networking (SDN) based load balancer which improves the efficiency of the network by distributing the traffic across multiple paths to optimize the efficiency of the network. Traditional load balancers are very expensive and inflexible. These SDN load balancers do not require costly hardware and can be programmed, which it makes it easier to implement user-defined algorithms and load balancing strategies. In this paper, we have proposed an efficient load balancing technique by considering different parameters to maintain the load efficiently using Open FlowSwitches connected to ONOS controller.


2021 ◽  
Vol 12 (1) ◽  
pp. 74-83
Author(s):  
Manjunatha S. ◽  
Suresh L.

Data center is a cost-effective infrastructure for storing large volumes of data and hosting large-scale service applications. Cloud computing service providers are rapidly deploying data centers across the world with a huge number of servers and switches. These data centers consume significant amounts of energy, contributing to high operational costs. Thus, optimizing the energy consumption of servers and networks in data centers can reduce operational costs. In a data center, power consumption is mainly due to servers, networking devices, and cooling systems, and an effective energy-saving strategy is to consolidate the computation and communication into a smaller number of servers and network devices and then power off as many unneeded servers and network devices as possible.


Sign in / Sign up

Export Citation Format

Share Document