Multicommodity max-flow min-cut theorems and their use in designing approximation algorithms
Journal of the ACM (JACM)
Communication Networks
On guaranteed smooth scheduling for input-queued switches
IEEE/ACM Transactions on Networking (TON)
Designing packet buffers for router linecards
IEEE/ACM Transactions on Networking (TON)
Deploying QoS for Cisco IP and Next Generation Networks: The Definitive Guide
Deploying QoS for Cisco IP and Next Generation Networks: The Definitive Guide
IEEE Journal on Selected Areas in Communications
Network traffic characteristics of data centers in the wild
IMC '10 Proceedings of the 10th ACM SIGCOMM conference on Internet measurement
Computer Architecture, Fifth Edition: A Quantitative Approach
Computer Architecture, Fifth Edition: A Quantitative Approach
The energy and emergy of the internet
Proceedings of the 10th ACM Workshop on Hot Topics in Networks
vSlicer: latency-aware virtual machine scheduling via differentiated-frequency CPU slicing
Proceedings of the 21st international symposium on High-Performance Parallel and Distributed Computing
Proceedings of the 21st international symposium on High-Performance Parallel and Distributed Computing
SpeQuloS: a QoS service for BoT applications using best effort distributed computing infrastructures
Proceedings of the 21st international symposium on High-Performance Parallel and Distributed Computing
Congestion control mechanisms and the best effort service model
IEEE Network: The Magazine of Global Internetworking
Hi-index | 0.00 |
This paper explores technologies to achieve low-latency energy-efficient communications in Global-Scale Cloud Computing systems. A global-scale cloud computing system linking 100 remote data-centers can interconnect potentially 5M servers, considerably larger than the size of traditional High-Performance-Computing (HPC) machines. Traditional HPC machines use tightly coupled processors and networks which rarely drop packets. In contrast, today's IP Internet is a relatively loosely-coupled Best-Effort network with poor latency and energy-efficiency guarantees, with relatively high packet loss rates. This paper explores the use of a recently-proposed Future-Internet network, which uses a QoS-aware router scheduling algorithm combined with a new IETF resource reservation signalling technology, to achieve improved latency and energy-efficiency in cloud computing systems. A Maximum-Flow Minimum-Energy routing algorithm is used to route high-capacity "trunks" between data-centers distributed over the continental USA, using a USA IP network topology. The communications between virtual machines in remote data-centers are aggregated and multiplexed onto the trunks, to achieve significantly improved energy-efficiency. According to theory and simulations, the large and variable queueing delays of traditional Best-Effort Internet links can be eliminated, and the latency over the cloud can be reduced to near-minimal values, i.e., the fiber latency. The maximum fiber latencies over the Sprint USA network are approx. 20 milliseconds, comparable to hard disk drive latencies, and multithreading in virtual machines can be used to hide these latencies. Furthermore, if existing dark-fiber over the continental network is activated, the bisection bandwidth available in a global-scale cloud computing system can rival that achievable in commercial HPC machines.