Architecture and early performance of the new IBM HPS fabric and adapter

  • Authors:
  • Rama K Govindaraju;Peter Hochschild;Don Grice;Kevin Gildea;Robert Blackmore;Carl A Bender;Chulho Kim;Piyush Chaudhary;Jason Goscinski;Jay Herring;Steven Martin;John Houston

  • Affiliations:
  • Server Development Lab, IBM, Poughkeepsie, NY;T J Watson Research Center, IBM Research, Hawthorne;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY;Server Development Lab, IBM, Poughkeepsie, NY

  • Venue:
  • HiPC'04 Proceedings of the 11th international conference on High Performance Computing
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we describe the architecture, design, and performance of the new cluster switch fabric and adapter called HPS (High Performance Switch) HPS delivers very low latency and very high bandwidth We demonstrate latency of less than 4.3us MPI library; 1.8GB/s of delivered unidirectional bandwidth and 2.9GB/s of bidirectional bandwidth between 2 MPI tasks running on 1.9GHz Power 4+ IH based nodes HPS also supports RDMA (remote direct memory access capability) A unique capability of RDMA over HPS is that reliable RDMA is supported over an underlying unreliable transport (unlike Infiniband and other RDMA transport protocols which depend on the underlying transport being reliable) We profile the performance of RDMA and its impact on striping for systems in which multiple network adapters are available to tasks of parallel jobs.