Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Neuro-Dynamic Programming
Scheduling dependent real-time activities
Scheduling dependent real-time activities
High Performance Computing Productivity Model Synthesis
International Journal of High Performance Computing Applications
Brief paper: Average cost temporal-difference learning
Automatica (Journal of IFAC)
Future Generation Computer Systems
Computational models and heuristic methods for Grid scheduling problems
Future Generation Computer Systems
Hi-index | 0.00 |
This paper addresses the problem of scheduling jobs in soft real-time systems, where the utility of completing each job decreases over time. We present a utility-based framework for making repeated scheduling decisions based on dynamically observed information about unscheduled jobs and system’s resources. This framework generalizes the standard scheduling problem to a resource-constrained environment, where resource allocation (RA) decisions (how many CPUs to allocate to each job) have to be made concurrently with the scheduling decisions (when to execute each job). Discrete-time Optimal Control theory is used to formulate the optimization problem of finding the scheduling/RA policy that maximizes the average utility per time step obtained from completed jobs. We propose a Reinforcement Learning (RL) architecture for solving the NP-hard Optimal Control problem in real time, and our experimental results demonstrate the feasibility and benefits of the proposed approach.