2022
DOI: 10.32604/cmc.2022.030803
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning-Based Job Shop Scheduling of燬mart燤anufacturing

Abstract: Industry 4.0 production environments and smart manufacturing systems integrate both the physical and decision-making aspects of manufacturing operations into autonomous and decentralized systems. One of the key aspects of these systems is a production planning, specifically, Scheduling operations on the machines. To cope with this problem, this paper proposed a Deep Reinforcement Learning with an Actor-Critic algorithm (DRLAC). We model the Job-Shop Scheduling Problem (JSSP) as a Markov Decision Process (MDP),… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 33 publications
(40 reference statements)
0
0
0
Order By: Relevance
“…The authors tested the solution in a real industrial case, and the algorithm turned out to provide a lower disassembly time than other algorithms, such as a GA, an improved discrete bee algorithm, and a dueling DQN. RL-Q-learning x [51] RL-AC algorithm x [52] RL-Q-learning x [53] RL-Q-learning + CTPNs x [54] RL-Q-learning x [55] MARL-Deep RL x [56] MARL-Deep RL x [57] MARL-SARSA x [58] RL-DQN x…”
Section: Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…The authors tested the solution in a real industrial case, and the algorithm turned out to provide a lower disassembly time than other algorithms, such as a GA, an improved discrete bee algorithm, and a dueling DQN. RL-Q-learning x [51] RL-AC algorithm x [52] RL-Q-learning x [53] RL-Q-learning + CTPNs x [54] RL-Q-learning x [55] MARL-Deep RL x [56] MARL-Deep RL x [57] MARL-SARSA x [58] RL-DQN x…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…A different type of problem and a different algorithm were presented by Elsayed, E.K, et al [51], who adopted the actor-critic (AC) network's training algorithm-based RL for achieving the optimal policy for the JSSP. The algorithm was tested in a real case, where scheduling was previously conducted following FIFO logic; the proposed algorithm achieved better results in terms of the makespan, by going from 97 UT to 60 UT.…”
Section: Reinforcement Learningmentioning
confidence: 99%