Simultaneous multithreading: maximizing on-chip parallelism
ISCA '95 Proceedings of the 22nd annual international symposium on Computer architecture
Effective distributed scheduling of parallel workloads
Proceedings of the 1996 ACM SIGMETRICS international conference on Measurement and modeling of computer systems
A closer look at coscheduling approaches for a network of workstations
Proceedings of the eleventh annual ACM symposium on Parallel algorithms and architectures
The impact of job memory requirements on gang-scheduling performance
ACM SIGMETRICS Performance Evaluation Review
Job scheduling in the presence of multiple resource requirements
SC '99 Proceedings of the 1999 ACM/IEEE conference on Supercomputing
A simulation-based study of scheduling mechanisms for a dynamic cluster environment
Proceedings of the 14th international conference on Supercomputing
Symbiotic jobscheduling for a simultaneous multithreaded processor
ASPLOS IX Proceedings of the ninth international conference on Architectural support for programming languages and operating systems
A Slowdown Model for Applications Executing on Time-Shared Clusters of Workstations
IEEE Transactions on Parallel and Distributed Systems
An infrastructure for efficient parallel job execution in Terascale computing environments
SC '98 Proceedings of the 1998 ACM/IEEE conference on Supercomputing
When the Herd Is Smart: Aggregate Behavior in the Selection of Job Request
IEEE Transactions on Parallel and Distributed Systems
Adaptive Scheduling under Memory Pressure on Multiprogrammed SMPs
IPDPS '02 Proceedings of the 16th International Parallel and Distributed Processing Symposium
IPPS '99/SPDP '99 Proceedings of the 13th International Symposium on Parallel Processing and the 10th Symposium on Parallel and Distributed Processing
A Historical Application Profiler for Use by Parallel Schedulers
IPPS '97 Proceedings of the Job Scheduling Strategies for Parallel Processing
Dynamic Coscheduling on Workstation Clusters
IPPS/SPDP '98 Proceedings of the Workshop on Job Scheduling Strategies for Parallel Processing
Characteristics of a Large Shared Memory Production Workload
JSSPP '01 Revised Papers from the 7th International Workshop on Job Scheduling Strategies for Parallel Processing
IPDPS '03 Proceedings of the 17th International Symposium on Parallel and Distributed Processing
Gang Scheduling with Memory Considerations
IPDPS '00 Proceedings of the 14th International Symposium on Parallel and Distributed Processing
IEEE Transactions on Parallel and Distributed Systems
The workload on parallel supercomputers: modeling the characteristics of rigid jobs
Journal of Parallel and Distributed Computing
Concurrency and Computation: Practice & Experience
Enhancements for hyper-threading technology in the operating system: seeking the optimal scheduling
WIESS'02 Proceedings of the 2nd conference on Industrial Experiences with Systems Software - Volume 2
ATOP-space and time adaptation for parallel and grid applications via flexible data partitioning
ARM '04 Proceedings of the 3rd workshop on Adaptive and reflective middleware
LOMARC: Lookahead Matchmaking for Multiresource Coscheduling on Hyperthreaded CPUs
IEEE Transactions on Parallel and Distributed Systems
Time and space adaptation for computational grids with the ATOP-Grid middleware
Future Generation Computer Systems
New challenges of parallel job scheduling
JSSPP'07 Proceedings of the 13th international conference on Job scheduling strategies for parallel processing
Impact of reservations on production job scheduling
JSSPP'07 Proceedings of the 13th international conference on Job scheduling strategies for parallel processing
ScoPred–scalable user-directed performance prediction using complexity modeling and historical data
JSSPP'05 Proceedings of the 11th international conference on Job Scheduling Strategies for Parallel Processing
Pitfalls in parallel job scheduling evaluation
JSSPP'05 Proceedings of the 11th international conference on Job Scheduling Strategies for Parallel Processing
Hi-index | 0.00 |
Job scheduling typically focuses on the CPU with little work existing to include I/O or memory. Time-shared execution provides the chance to hide I/O and long-communication latencies though potentially creating a memory conflict. We consider two different cases: standard local CPU scheduling and coscheduling on hyperthreaded CPUs. The latter supports coscheduling without any context switches and provides additional options for CPU-internal resource sharing. We present an approach that includes all possible resources into the schedule optimization and improves utilization by coscheduling two jobs if feasible. Our LOMARC approach partially reorders the queue by lookahead to increase the potential to find good matches. In simulations based on the workload model of [12], we have obtained improvements of about 50% in both response times and relative bounded response times on hyperthreaded CPUs (i.e. cut times by half) and of about 25% on standard CPUs for our LOMARC scheduling approach.