An novel cloud task scheduling framework using hierarchical deep reinforcement learning for cloud computing.

Journal: PloS one
Published Date:

Abstract

With the increasing popularity of cloud computing services, their large and dynamic load characteristics have rendered task scheduling an NP-complete problem.To address the problem of large-scale task scheduling in a cloud computing environment, this paper proposes a novel cloud task scheduling framework using hierarchical deep reinforcement learning (DRL) to address the challenges of large-scale task scheduling in cloud computing. The framework defines a set of virtual machines (VMs) as a VM cluster and employs hierarchical scheduling to allocate tasks first to the cluster and then to individual VMs. The scheduler, designed using DRL, adapts to dynamic changes in the cloud environments by continuously learning and updating network parameters. Experiments demonstrate that it skillfully balances cost and performance. In low-load situations, costs are reduced by using low-cost nodes within the Service Level Agreement (SLA) range; in high-load situations, resource utilization is improved through load balancing. Compared with classical heuristic algorithms, it effectively optimizes load balancing, cost, and overdue time, achieving a 10% overall improvement. The experimental results demonstrate that this approach effectively balances cost and performance, optimizing objectives such as load balance, cost, and overdue time. One potential shortcoming of the proposed hierarchical deep reinforcement learning (DRL) framework for cloud task scheduling is its complexity and computational overhead. Implementing and maintaining a DRL-based scheduler requires significant computational resources and expertise in machine learning. There are still shortcomings in the method used in this study. First, the continuous learning and updating of network parameters might introduce latency, which could impact real-time task scheduling efficiency. Furthermore, the framework's performance heavily depends on the quality and quantity of training data, which might be challenging to obtain and maintain in a dynamic cloud environment.

Authors

  • Delong Cui
    College of Computer and Electronic Information, Guangdong University of Petrochemical Technology, Maoming 525000, China.
  • Zhiping Peng
    School of Automation, Guangdong University of Petrochemical Technology, Maoming 525000, PR China. Electronic address: pengzp@gdupt.edu.cn.
  • Kaibin Li
    College of Electronic Information Engineering, Guangdong University of Petrochemical Technology, Maoming, China.
  • Qirui Li
    College of Computer and Electronic Information, Guangdong University of Petrochemical Technology, Maoming 525000, China.
  • Jieguang He
    College of Computer and Electronic Information, Guangdong University of Petrochemical Technology, Maoming 525000, China.
  • Xiangwu Deng
    College of Engineering, South China Agricultural University, Guangzhou, China.