2021
DOI: 10.2507/ijsimm20-2-co7
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Reinforcement Learning Based Solution for Flexible Job Shop Scheduling Problem

Abstract: Flexible job shop Scheduling problem (FJSP) is a classic problem in combinatorial optimization and a very common form of organization in a real production environment. Traditional approaches for FJSP are ill-suited to deal with complex and changeable production environments. Based on 3D disjunctive graph dispatching, this work proposes an end-to-end deep reinforcement learning (DRL) framework. In this framework, a modified pointer network, which consists of an encoder and a decoder, is adopted to encode the op… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 32 publications
(8 citation statements)
references
References 13 publications
0
8
0
Order By: Relevance
“…where A t is the advantage function computed as the difference between the discounted sum of rewards and the baseline estimate at the state as shown in (21), r t (θ) is the ratio of the new policy to the old policy as described in (22), and is the clipping ratio.…”
Section: Drl Algorithm For Policy Training Since Thementioning
confidence: 99%
See 1 more Smart Citation
“…where A t is the advantage function computed as the difference between the discounted sum of rewards and the baseline estimate at the state as shown in (21), r t (θ) is the ratio of the new policy to the old policy as described in (22), and is the clipping ratio.…”
Section: Drl Algorithm For Policy Training Since Thementioning
confidence: 99%
“…Recently the use of DRL methods for solving scheduling problems is gaining attention and, promising results are being obtained in job shop scheduling problem (JSSP) [21], flexible job shop scheduling problem (FJSP) [22], and open shop scheduling problem (OSSP) [23]. In Ref.…”
Section: Review Of Literaturementioning
confidence: 99%
“…In today's complex and varied production processes, dynamic events such as machine breakdown or the change of the processing time and machine order of jobs are inevitable to be considered, necessitating the remarkable results on various combinatorial optimization problems such as traveling salesman problem (TSP) [12], the vehicle routing problem (VRP) [13] and JSP [14]. Existing research has shown that using reinforcement learning to solve DJSP has at least four advantages: 1) RL doesn't require the complete mathematical model or the large labeled datasets of the scheduling environment, but can learn from the interaction with the environment and store the learned knowledge to achieve "offine learning and online application" [15]. 2) Unlike the exiting approaches that have to reschedule the jobs when faced dynamic events, RL can adjust the learning strategy automatically and achieve adaptive scheduling [16].…”
Section: Introductionmentioning
confidence: 99%
“…Machines with intelligent agents evaluate the priorities of jobs and distribute them through negotiation. Han et al [27] proposed an end-to-end DRL framework based on 3D disjunctive graph dispatching. They improved the pointer network and trained the policy with 20 static features and 24 dynamic features that described the full picture of scheduling problem.…”
Section: Introductionmentioning
confidence: 99%