An efficient deep reinforcement learning environment for flexible job-shop scheduling
By: Xinquan Wu , Xuefeng Yan , Mingqiang Wei and more
Potential Business Impact:
Makes factory jobs finish faster and better.
The Flexible Job-shop Scheduling Problem (FJSP) is a classical combinatorial optimization problem that has a wide-range of applications in the real world. In order to generate fast and accurate scheduling solutions for FJSP, various deep reinforcement learning (DRL) scheduling methods have been developed. However, these methods are mainly focused on the design of DRL scheduling Agent, overlooking the modeling of DRL environment. This paper presents a simple chronological DRL environment for FJSP based on discrete event simulation and an end-to-end DRL scheduling model is proposed based on the proximal policy optimization (PPO). Furthermore, a short novel state representation of FJSP is proposed based on two state variables in the scheduling environment and a novel comprehensible reward function is designed based on the scheduling area of machines. Experimental results on public benchmark instances show that the performance of simple priority dispatching rules (PDR) is improved in our scheduling environment and our DRL scheduling model obtains competing performance compared with OR-Tools, meta-heuristic, DRL and PDR scheduling methods.
Similar Papers
Adaptive Bias Generalized Rollout Policy Adaptation on the Flexible Job-Shop Scheduling Problem
Artificial Intelligence
Makes factory jobs finish faster.
A Production Scheduling Framework for Reinforcement Learning Under Real-World Constraints
Machine Learning (CS)
Helps factories make things faster and better.
A Benchmarking Environment for Worker Flexibility in Flexible Job Shop Scheduling Problems
Neural and Evolutionary Computing
Helps factories make things faster and cheaper.