Publication | Open Access
Multi-Objective Workflow Scheduling With Deep-Q-Network-Based Multi-Agent Reinforcement Learning
277
Citations
31
References
2019
Year
Cloud computing enables large‑scale workflow execution, yet optimal scheduling for multiple conflicting objectives remains inadequately addressed, with existing methods limited by expert‑dependent encoding that hampers performance. This study applies a deep‑Q‑network in a multi‑agent reinforcement learning framework to schedule multiple workflows on IaaS clouds. The authors formulate scheduling as a Markov game whose state encodes workflow counts and heterogeneous VMs, rewards combine makespan and cost, and the game seeks a correlated equilibrium without prior knowledge, converging in dynamic real‑time and validated on scientific workflow templates and Amazon EC2. Experimental results show the proposed approach outperforms traditional multi‑objective scheduling algorithms such as NSGA‑II, MOPSO, and game‑theoretic greedy methods in generating more optimal plans.
Cloud Computing provides an effective platform for executing large-scale and complex workflow applications with a pay-as-you-go model. Nevertheless, various challenges, especially its optimal scheduling for multiple conflicting objectives, are yet to be addressed properly. The existing multi-objective workflow scheduling approaches are still limited in many ways, e.g., encoding is restricted by prior experts' knowledge when handling a dynamic real-time problem, which strongly influences the performance of scheduling. In this paper, we apply a deep-Q-network model in a multi-agent reinforcement learning setting to guide the scheduling of multi-workflows over infrastructure-as-a-service clouds. To optimize multi-workflow completion time and user's cost, we consider a Markov game model, which takes the number of workflow applications and heterogeneous virtual machines as state input and the maximum completion time and cost as rewards. The game model is capable of seeking for correlated equilibrium between make-span and cost criteria without prior experts' knowledge and converges to the correlated equilibrium policy in a dynamic real-time environment. To validate our proposed approach, we conduct extensive case studies based on multiple well-known scientific workflow templates and Amazon EC2 cloud. The experimental results clearly suggest that our proposed approach outperforms traditional ones, e.g., non-dominated sorting genetic algorithm-II, multi-objective particle swarm optimization, and game-theoretic-based greedy algorithms, in terms of optimality of scheduling plans generated.
| Year | Citations | |
|---|---|---|
2002 | 46.1K | |
2004 | 4.3K | |
2016 | 1.1K | |
2014 | 444 | |
2016 | 350 | |
2018 | 349 | |
2014 | 273 | |
2017 | 253 | |
2018 | 251 | |
2018 | 206 |
Page 1
Page 1