Gprof: A call graph execution profiler
SIGPLAN '82 Proceedings of the 1982 SIGPLAN symposium on Compiler construction
Xen and the art of virtualization
SOSP '03 Proceedings of the nineteenth ACM symposium on Operating systems principles
Multi-dimensional storage virtualization
Proceedings of the joint international conference on Measurement and modeling of computer systems
Diagnosing performance overheads in the xen virtual machine environment
Proceedings of the 1st ACM/USENIX international conference on Virtual execution environments
Buttress: A Toolkit for Flexible and High Fidelity I/O Benchmarking
FAST '04 Proceedings of the 3rd USENIX Conference on File and Storage Technologies
I/O system performance debugging using model-driven anomaly characterization
FAST'05 Proceedings of the 4th conference on USENIX Conference on File and Storage Technologies - Volume 4
Accurate and efficient replaying of file system traces
FAST'05 Proceedings of the 4th conference on USENIX Conference on File and Storage Technologies - Volume 4
lmbench: portable tools for performance analysis
ATEC '96 Proceedings of the 1996 annual conference on USENIX Annual Technical Conference
Hi-index | 0.00 |
As distributed systems increase in both popularity and scale, it becomes increasingly important to understand as well as to systematically identify performance anomalies and potential opportunities for optimization. However, large scale distributed systems are often complex and non-deterministic due to hardware and software heterogeneity and configurable runtime options that may boost or diminish performance. It is therefore important to be able to disseminate and present the information gleaned from a local system under a common evaluation methodology so that such efforts can be valuable in one environment and provide general guidelines for other environments. Evaluation methodologies can conveniently be encapsulated inside of a common analysis framework that serves as an outer layer upon which appropriate experimental design and relevant workloads (benchmarking and profiling applications) can be supported. In this paper we present ExPerT, an Ex tensible Per formance T oolkit. ExPerT defines a flexible framework from which a set of benchmarking, tracing, and profiling applications can be correlated together in a unified interface. The framework consists primarily of two parts: an extensible module for profiling and benchmarking support, and a unified data discovery tool for information gathering and parsing. We include a case study of disk I/O performance in virtualized distributed environments which demonstrates the flexibility of our framework for selecting benchmark suite, creating experimental design, and performing analysis.