2022
DOI: 10.36227/techrxiv.20324070
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

You Only Train Once: A highly generalizable reinforcement learning method for dynamic job shop scheduling problem

Abstract: <p>Research in artificial intelligence demonstrates the applicability and flexibility of the reinforcement learning (RL) technique for the dynamic job shop scheduling problem (DJSP). However, the RL-based method will always overfit to the training environment and cannot generalize well to novel unseen situations at deployment time, which is unacceptable in real-world production. For this reason, this paper proposes a highly generalizable reinforcement learning framework named Train Once For All (TOFA) fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 24 publications
0
3
0
Order By: Relevance
“…Some methods start considering various dynamic events to make the scheduler more robust to these disturbances. Zeng et al [94] consider machine breakdown and different order requirements in JSSP, where a machine can break or the configuration of jobs may change. They formulate the DJSSP as an MDP with the disjunctive graphs as the states and a set of PDRs as the action space.…”
Section: Learning To Solve Jssp With Dynamic Eventsmentioning
confidence: 99%
“…Some methods start considering various dynamic events to make the scheduler more robust to these disturbances. Zeng et al [94] consider machine breakdown and different order requirements in JSSP, where a machine can break or the configuration of jobs may change. They formulate the DJSSP as an MDP with the disjunctive graphs as the states and a set of PDRs as the action space.…”
Section: Learning To Solve Jssp With Dynamic Eventsmentioning
confidence: 99%
“…The state of the manufacturing environment can be inaccurately and incompletely expressed due to artificial factors. Action space is mainly designed as priority rules 32,33 or parameter optimization, 34 which don't meet the desired execution efficiency in action exploration. In addition, some work even generalizes the model developed in the static environment to an uncertain resource environment, 35 lacking the learning process in the dynamic environment.…”
Section: Rl-based Dynamical Schedulingmentioning
confidence: 99%
“…Applied research often considers an additional dimension in the problem formulation inspired by real-world use-cases, such as stochasticity [13,14], machine flexibility [15][16][17], dynamic job releases [18], machine failures [19] or multi-objective optimization criteria [20,21]. These studies show the general feasibility of DRL to learn, but are typically not very competitive with expert systems.…”
Section: Deep Reinforcement Learning For Job Shop Scheduling Problemsmentioning
confidence: 99%