MapReduce: simplified data processing on large clusters
Communications of the ACM - 50th anniversary issue: 1958 - 2008
Dynamic proportional share scheduling in Hadoop
JSSPP'10 Proceedings of the 15th international conference on Job scheduling strategies for parallel processing
Scheduling Hadoop Jobs to Meet Deadlines
CLOUDCOM '10 Proceedings of the 2010 IEEE Second International Conference on Cloud Computing Technology and Science
A load-aware scheduler for MapReduce framework in heterogeneous cloud environments
Proceedings of the 2011 ACM Symposium on Applied Computing
Scheduling Mixed Real-Time and Non-real-Time Applications in MapReduce Environment
ICPADS '11 Proceedings of the 2011 IEEE 17th International Conference on Parallel and Distributed Systems
Hi-index | 0.00 |
MapReduce is a programming model and an associated implementation for processing and generating large data sets. Providing MapReduce as a service is the development future trend. By leveraging the game theory, this paper proposes a scheduling algorithm to deal with the competition for resources between multiple jobs in MapReduce. Firstly, we present a model that could estimate job executing time, and then a utility function of job and an optimization objective are brought forward; thirdly, we present a game model to solve the optimization problem. The proof and the solution are also present. Finally, we implement the algorithm and experiment it in a hadoop cluster. The result shows the present algorithm could schedule jobs rational.