Abstract

In the past decades, many optimization methods have been devised and applied to job shop scheduling problem (JSSP) to find the optimal solution. Many methods assumed that the scheduling results were applied to static environments, but the whole environments in the real world are always dynamic. Moreover, many unexpected events such as machine breakdowns and material problems may be present to adversely affect the initial job scheduling. This work views JSSP as a sequential decision making problem and proposes to use deep reinforcement learning to cope with this problem. The combination of deep learning and reinforcement learning avoids handcraft features as used in traditional reinforcement learning, and it is expected that the combination will make the whole learning phase more efficient. Our proposed model comprises actor network and critic network, both including convolution layers and fully connected layer. Actor network agent learns how to behave in different situations, while critic network helps agent evaluate the value of statement then return to actor network. This work proposes a parallel training method, combining asynchronous update as well as deep deterministic policy gradient (DDPG), to train the model. The whole network is trained with parallel training on a multi-agent environment and different simple dispatching rules are considered as actions. We evaluate our proposed model on more than ten instances that are present in a famous benchmark problem library - OR library. The evaluation results indicate that our method is comparative in static JSSP benchmark problems, and achieves a good balance between makespan and execution time in dynamic environments. Scheduling score of our method is 91.12% in static JSSP benchmark problems, and 80.78% in dynamic environments.

Highlights

  • Scheduling can be viewed as a decision making process, in which the scheduling arranges the tasks that we should process sequentially

  • We further investigate the impact of different rewards on our proposed method, and we summarize the performances for these two rewards on all the tasks in Fig. 6, which indicates that using combined reward in our deep reinforcement learning (DRL) model normally leads to better results on makespan as compared with using fixed reward in the end of training episode

  • In this work we propose an actor–critic architecture with DRL to cope with classic job shop problems

Read more

Summary

Introduction

Scheduling can be viewed as a decision making process, in which the scheduling arranges the tasks that we should process sequentially. Scheduling is an essential task in many application domains, including manufacturing [13], [43], [44], network [40] and supply chains [12], [36]. Wang et al [36] considered production and distribution in supply chain management, and focused on the permutation flow shop scheduling problem, which addressed the problem of batch delivery to multiple customers. Zheng and Wang [44] focused on a resource constrained unrelated parallel machine green manufacturing scheduling problem, which aimed to minimize the makespan and the total carbon emission. Mathematical programming is a popular approach to deal with the scheduling problem by formulating it as an optimization problem with constraints, and using optimization techniques to seek the optimal solution

Objectives
Methods
Findings
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call