Processor-sharing queues: some progress in analysis
Queueing Systems: Theory and Applications
Models for studying concurrency control performance: alternatives and implications
SIGMOD '85 Proceedings of the 1985 ACM SIGMOD international conference on Management of data
Time-shared Systems: a theoretical treatment
Journal of the ACM (JACM)
Waiting Time Distributions for Processor-Sharing Systems
Journal of the ACM (JACM)
Communications of the ACM
SEDA: an architecture for well-conditioned, scalable internet services
SOSP '01 Proceedings of the eighteenth ACM symposium on Operating systems principles
Sojourn time asymptotics in the M/G/1 processor sharing queue
Queueing Systems: Theory and Applications
Adaptive Load Control in Transaction Processing Systems
VLDB '91 Proceedings of the 17th International Conference on Very Large Data Bases
A method for transparent admission control and request scheduling in e-commerce web sites
Proceedings of the 13th international conference on World Wide Web
SIGMETRICS '82 Proceedings of the 1982 ACM SIGMETRICS conference on Measurement and modeling of computer systems
How to Determine a Good Multi-Programming Level for External Scheduling
ICDE '06 Proceedings of the 22nd International Conference on Data Engineering
Introduction to special issue on queueing models for fair resource sharing
Queueing Systems: Theory and Applications
Steady state approximations of limited processor sharing queues in heavy traffic
Queueing Systems: Theory and Applications
Tail-robust scheduling via limited processor sharing
Performance Evaluation
CODES+ISSS '11 Proceedings of the seventh IEEE/ACM/IFIP international conference on Hardware/software codesign and system synthesis
Delay tails in MapReduce scheduling
Proceedings of the 12th ACM SIGMETRICS/PERFORMANCE joint international conference on Measurement and Modeling of Computer Systems
Hi-index | 0.00 |
We consider the problem of admission control in resource sharing systems, such as web servers and transaction processing systems, when the job size distribution has high variability, with the aim of minimizing the mean response time. It is well known that in such resource sharing systems, as the number of tasks concurrently sharing the resource is increased, the server throughput initially increases, due to more efficient utilization of resources, but starts falling beyond a certain point, due to resource contention and thrashing. Most admission control mechanisms solve this problem by imposing a fixed upper bound on the number of concurrent transactions allowed into the system, called the Multi-Programming-Limit (MPL), and making the arrivals which find the server full queue up. Almost always, the MPL is chosen to be the point that maximizes server efficiency. In this paper we abstract such resource sharing systems as a Processor Sharing (PS) server with state-dependent service rate and a First-Come-First-Served (FCFS) queue, and we analyze the performance of this model from a queueing theoretic perspective. We start by showing that, counter to the common wisdom, the peak efficiency point is not always optimal for minimizing the mean response time. Instead, significant performance gains can be obtained by running the system at less than the peak efficiency. We provide a simple expression for the static MPL that achieves near-optimal mean response time for general distributions. Next we present two traffic-oblivious dynamic admission control policies that adjust the MPL based on the instantaneous queue length while also taking into account the variability of the job size distribution. The structure of our admission control policies is a mixture of fluid control when the number of jobs in the system is high, with a stochastic component when the system is near-empty. We show via simulations that our dynamic policies are much more robust to unknown traffic intensities and burstiness in the arrival process than imposing a static MPL.