2022
DOI: 10.1016/j.rcim.2022.102412
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic job shop scheduling based on deep reinforcement learning for multi-agent manufacturing systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
3

Relationship

3
6

Authors

Journals

citations
Cited by 86 publications
(30 citation statements)
references
References 40 publications
0
18
0
Order By: Relevance
“…(iii) It can be found that the DRL significantly outperforms all scheduling rules when trained on small-scale instances and generalized on large-scale instances, indicating that the method proposed in this study is effective when dealing with highdimensional input space; and for the whole learning process, DMU is the data used for testing, and it can be seen from the experimental data that the method proposed in this study can effectively learn to generate better for invisible instances solutions. (iv) Tested with the same parameters, the PPO algorithm [44] performs better on instances than DQN [41] and DDPG [58] and performs about the same as the metaheuristic on instances with a relatively small total number of JXMs but for larger instances, the performance of the method proposed in this study is significantly better. However, overall, regardless of the method used, the ability to solve large-scale problems is worse than the ability to solve small-scale problems, and the training error increases as the scale increases in comparison to DRL.…”
Section: Resultsmentioning
confidence: 84%
See 1 more Smart Citation
“…(iii) It can be found that the DRL significantly outperforms all scheduling rules when trained on small-scale instances and generalized on large-scale instances, indicating that the method proposed in this study is effective when dealing with highdimensional input space; and for the whole learning process, DMU is the data used for testing, and it can be seen from the experimental data that the method proposed in this study can effectively learn to generate better for invisible instances solutions. (iv) Tested with the same parameters, the PPO algorithm [44] performs better on instances than DQN [41] and DDPG [58] and performs about the same as the metaheuristic on instances with a relatively small total number of JXMs but for larger instances, the performance of the method proposed in this study is significantly better. However, overall, regardless of the method used, the ability to solve large-scale problems is worse than the ability to solve small-scale problems, and the training error increases as the scale increases in comparison to DRL.…”
Section: Resultsmentioning
confidence: 84%
“…Liu et al [43] introduced an actor-critic deep reinforcement learning approach grounded on scheduling rules to determine actions. Zhang [44] presented an AI scheduler adaptive learning strategy rooted in the proximal policy optimization (PPO) algorithm to enhance decision-making capabilities amidst order and resource disruptions. Han [45] presented a deep reinforcement learning (DRL) framework that leverages analytical graph scheduling to navigate the complex and dynamic production environment inherent to dynamic job shop scheduling problems.…”
Section: Dynamic Job Shop Scheduling Based On Artificial Intelligence...mentioning
confidence: 99%
“…Cai et al [25] developed a large-scale and dynamic job-shop simulation platform. Zhang et al [26] proposed a multi-intelligence architecture system integrating self-organizing negotiation and self-learning strategies.…”
Section: Intelligent Production Scheduling Methodsmentioning
confidence: 99%
“…Less works consider mobility with applications using AGVs or mobile robots, i.e., [111] and in some cases, they are still in a very conceptual implementation stage [90]. In [107], it is presented an approach based on multi-agent negotiation that showcases production self-organization in a smart workshop, AGVs are in charge of the transportation of material from the warehouse to production cells. Fig.…”
Section: B Automation Levelmentioning
confidence: 99%