Measurements of a distributed file system

  • Authors:
  • Mary G. Baker;John H. Hartman;Michael D. Kupfer;Ken W. Shirriff;John K. Ousterhout

  • Affiliations:
  • Computer Science Division, Electrical Engineering and Computer Sciences, University of California, Berkeley, CA;Computer Science Division, Electrical Engineering and Computer Sciences, University of California, Berkeley, CA;Computer Science Division, Electrical Engineering and Computer Sciences, University of California, Berkeley, CA;Computer Science Division, Electrical Engineering and Computer Sciences, University of California, Berkeley, CA;Computer Science Division, Electrical Engineering and Computer Sciences, University of California, Berkeley, CA

  • Venue:
  • SOSP '91 Proceedings of the thirteenth ACM symposium on Operating systems principles
  • Year:
  • 1991

Quantified Score

Hi-index 0.02

Visualization

Abstract

We analyzed the user-level file access patterns and caching behavior of the Sprite distributed file system. The first part of our analysis repeated a study done in 1985 of the: BSD UNIX file system. We found that file throughput has increased by a factor of 20 to an average of 8 Kbytes per second per active user over 10-minute intervals, and that the use of process migration for load sharing increased burst rates by another factor of six. Also, many more very large (multi-megabyte) files are in use today than in 1985. The second part of our analysis measured the behavior of Sprite's main-memory file caches. Client-level caches average about 7 Mbytes in size (about one-quarter to one-third of main memory) and filter out about 50% of the traffic between clients and servers. 35% of the remaining server traffic is caused by paging, even on workstations with large memories. We found that client cache consistency is needed to prevent stale data errors, but that it is not invoked often enough to degrade overall system performance.