Proceedings of the 1989 ACM/IEEE conference on Supercomputing
Interaction cost and shotgun profiling
ACM Transactions on Architecture and Code Optimization (TACO)
Automatic measurement of memory hierarchy parameters
SIGMETRICS '05 Proceedings of the 2005 ACM SIGMETRICS international conference on Measurement and modeling of computer systems
Investigating Cache Parameters of x86 Family Processors
Proceedings of the 2009 SPEC Benchmark Workshop on Computer Performance Evaluation and Benchmarking
Producing wrong data without doing anything obviously wrong!
Proceedings of the 14th international conference on Architectural support for programming languages and operating systems
When Misses Differ: Investigating Impact of Cache Misses on Observed Performance
ICPADS '09 Proceedings of the 2009 15th International Conference on Parallel and Distributed Systems
Precise regression benchmarking with random effects: improving mono benchmark results
EPEW'06 Proceedings of the Third European conference on Formal Methods and Stochastic Models for Performance Evaluation
Can linear approximation improve performance prediction ?
EPEW'11 Proceedings of the 8th European conference on Computer Performance Engineering
Hi-index | 0.00 |
The memory subsystems of contemporary computer architectures are increasingly complex --- in fact, so much so that it becomes difficult to estimate the performance impact of many coding constructs, and some long known coding patterns are even discovered to be principally wrong. In contrast, many researchers still reason about algorithmic complexity in simple terms, where memory operations are sequential and of equal cost. The goal of this talk is to give an overview of some memory subsystem features that violate this assumption significantly, with the ambition to motivate development of algorithms tailored to contemporary computer architectures.