Experience distributing objects in an SMMP OS

  • Authors:
  • Jonathan Appavoo;Dilma Da Silva;Orran Krieger;Marc Auslander;Michal Ostrowski;Bryan Rosenburg;Amos Waterland;Robert W. Wisniewski;Jimi Xenidis;Michael Stumm;Livio Soares

  • Affiliations:
  • IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;IBM T.J. Watson Research Center;Dept. of Electrical and Computer Engineering, University of Toronto;Dept. of Electrical and Computer Engineering, University of Toronto

  • Venue:
  • ACM Transactions on Computer Systems (TOCS)
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Designing and implementing system software so that it scales well on shared-memory multiprocessors (SMMPs) has proven to be surprisingly challenging. To improve scalability, most designers to date have focused on concurrency by iteratively eliminating the need for locks and reducing lock contention. However, our experience indicates that locality is just as, if not more, important and that focusing on locality ultimately leads to a more scalable system. In this paper, we describe a methodology and a framework for constructing system software structured for locality, exploiting techniques similar to those used in distributed systems. Specifically, we found two techniques to be effective in improving scalability of SMMP operating systems: (i) an object-oriented structure that minimizes sharing by providing a natural mapping from independent requests to independent code paths and data structures, and (ii) the selective partitioning, distribution, and replication of object implementations in order to improve locality. We describe concrete examples of distributed objects and our experience implementing them. We demonstrate that the distributed implementations improve the scalability of operating-system-intensive parallel workloads.