Score: 0

An efficient deep reinforcement learning environment for flexible job-shop scheduling

Published: September 7, 2025 | arXiv ID: 2509.07019v1

By: Xinquan Wu , Xuefeng Yan , Mingqiang Wei and more

Potential Business Impact:

Makes factory jobs finish faster and better.

Business Areas:
Scheduling Information Technology, Software

The Flexible Job-shop Scheduling Problem (FJSP) is a classical combinatorial optimization problem that has a wide-range of applications in the real world. In order to generate fast and accurate scheduling solutions for FJSP, various deep reinforcement learning (DRL) scheduling methods have been developed. However, these methods are mainly focused on the design of DRL scheduling Agent, overlooking the modeling of DRL environment. This paper presents a simple chronological DRL environment for FJSP based on discrete event simulation and an end-to-end DRL scheduling model is proposed based on the proximal policy optimization (PPO). Furthermore, a short novel state representation of FJSP is proposed based on two state variables in the scheduling environment and a novel comprehensible reward function is designed based on the scheduling area of machines. Experimental results on public benchmark instances show that the performance of simple priority dispatching rules (PDR) is improved in our scheduling environment and our DRL scheduling model obtains competing performance compared with OR-Tools, meta-heuristic, DRL and PDR scheduling methods.

Country of Origin
šŸ‡ØšŸ‡³ China

Page Count
13 pages

Category
Computer Science:
Machine Learning (CS)