2021
DOI: 10.1080/00207543.2020.1870013
|View full text |Cite
|
Sign up to set email alerts
|

Learning to schedule job-shop problems: representation and policy learning using graph neural network and reinforcement learning

Abstract: We propose a framework to learn to schedule a job-shop problem (JSSP) using a graph neural network (GNN) and reinforcement learning (RL). We formulate the scheduling process of JSSP as a sequential decision-making problem with graph representation of the state to consider the structure of JSSP. In solving the formulated problem, the proposed framework employs a GNN to learn that node features that embed the spatial structure of the JSSP represented as a graph (representation learning) and derive the optimum sc… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 153 publications
(64 citation statements)
references
References 41 publications
0
40
0
Order By: Relevance
“…It was able to outperform conventional heuristics such as shortest setup or processing time and reduced total makespan and computation times after training. A further increase in generalisation was reached by Park et al (2021) by applying a graph neural network (GNN). The GNN learned the basic spatial structure of the problem in form of a graph that could be transferred to new problems and adapted its mapped policy.…”
Section: Production Schedulingmentioning
confidence: 99%
“…It was able to outperform conventional heuristics such as shortest setup or processing time and reduced total makespan and computation times after training. A further increase in generalisation was reached by Park et al (2021) by applying a graph neural network (GNN). The GNN learned the basic spatial structure of the problem in form of a graph that could be transferred to new problems and adapted its mapped policy.…”
Section: Production Schedulingmentioning
confidence: 99%
“…For example, [9,23] have proposed to learn scheduling policy for each agent and hence, it requires an additional training to solve JSPs with a different number of agents from the training cases. Recently, [30,45] have proposed to learn a shared scheduling policy for all agents while utilizing the disjunctive graph representation of JSP. Unlike these methods utilizing well-designed dense reward, we directly use the makespan reward to train a policy.…”
Section: Related Workmentioning
confidence: 99%
“…We evaluate the scheduling performance to verify ScheduleNet's generalization capacity to unseen JSP distributions on the Taillard's 80 dataset [39]. We compare against two deep RL baselines [30,45], as well as the mentioned heuristics (MOR, FIFO, and SPT).…”
Section: Jsp Experimentsmentioning
confidence: 99%
See 1 more Smart Citation
“…There exist several works on applying DRL for scheduling tasks in the various domains such as cloud computing, networks, and manufacturing systems [8]- [11], [15]- [18]. DeepRM [8] was the first attempt to learn a scheduling policy systematically through DRL.…”
Section: Namementioning
confidence: 99%