2023
DOI: 10.1109/tits.2022.3216203
|View full text |Cite
|
Sign up to set email alerts
|

Distributed Signal Control of Arterial Corridors Using Multi-Agent Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 29 publications
0
1
0
Order By: Relevance
“…This section evaluates the performance of OAS Deep Q-Learning is compared with those of two categories of the baseline methods: traditional methods (including MaxPressure [ 49 ] and FixedTime [ 50 ]) and the Multi-Agent Deep Reinforcement Learning (MARL) [ 51 ], Meta Variationally Intrinsic Motivated Reinforcement Learning (MetaVIMRL) [ 52 ], and Cooperative Multi-Agent Deep Q-Network (CMDQN) [ 53 ] based on the training results. For fair comparisons, the parameters and simulation conditions of the traditional methods and the DQL methods [ 51 , 52 , 53 ] are the same. Table 9 presents the parameters of OAS Deep Q-Learning with reinforcement learning network as mentioned in this paper.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This section evaluates the performance of OAS Deep Q-Learning is compared with those of two categories of the baseline methods: traditional methods (including MaxPressure [ 49 ] and FixedTime [ 50 ]) and the Multi-Agent Deep Reinforcement Learning (MARL) [ 51 ], Meta Variationally Intrinsic Motivated Reinforcement Learning (MetaVIMRL) [ 52 ], and Cooperative Multi-Agent Deep Q-Network (CMDQN) [ 53 ] based on the training results. For fair comparisons, the parameters and simulation conditions of the traditional methods and the DQL methods [ 51 , 52 , 53 ] are the same. Table 9 presents the parameters of OAS Deep Q-Learning with reinforcement learning network as mentioned in this paper.…”
Section: Resultsmentioning
confidence: 99%
“…To test the efficacy of the model proposed in this paper, the traffic flow data at the intersections 1-3 are presented in Tables 6-8. This section evaluates the performance of OAS Deep Q-Learning is compared with those of two categories of the baseline methods: traditional methods (including MaxPressure [49] and FixedTime [50]) and the Multi-Agent Deep Reinforcement Learning (MARL) [51], Meta Variationally Intrinsic Motivated Reinforcement Learning (MetaVIMRL) [52], and Cooperative Multi-Agent Deep Q-Network (CMDQN) [53] based on the training results. For fair comparisons, the parameters and simulation conditions of the traditional methods and the DQL methods [51][52][53] are the same.…”
Section: Collection Of Data On Arterial Intersectionsmentioning
confidence: 99%
“…When primary vehicles are close to an intersection, coordinated phases can turn green in time without constraints; however, there is a large safety risk owing to deviations from the real-world traffic environment. The other type [1,22,23] considers these constraints and can achieve similar vehicle progression through explicit signal coordination. The weights of the neural networks embed the maintenance of vehicle progression and improvement of system performance.…”
Section: Rl Methods For Signal Coordinationmentioning
confidence: 99%
“…To achieve this, the primary vehicles should be delayed as less as possible by the downstream signals. In real-world arterials, the phase sequence at an intersection is typically fixed over time for safety reasons [1]. This means that every signal must display green in a cyclical and fixed sequence, and that the primary vehicles arriving at the downstream intersection cyclically cannot be served in a timely manner.…”
Section: Introductionmentioning
confidence: 99%