2021
DOI: 10.1016/j.trc.2021.103059
|View full text |Cite
|
Sign up to set email alerts
|

Network-wide traffic signal control optimization using a multi-agent deep reinforcement learning

Abstract: A multi-agent reinforcement learning for adaptive traffic signal control optimization.• Consider control unit at intersection as agent that can communicate with others through knowledge-sharing protocol.• Proposed algorithm achieves consistent improvements over baselines on both simulated and real-world data.

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
44
0
6

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 83 publications
(50 citation statements)
references
References 38 publications
0
44
0
6
Order By: Relevance
“…One-hundred forty-four fully fledged research articles were finally selected based on the following inclusion criteria: most relevant, most cited, and most recent. For traffic state forecasting, in terms of performances, the GAN-based methods and also hybrid approaches showed better performance on state-of-the-art datasets, i.e., PeMS (Li et al [154], Zhang et al [95]). For intersection signal control, DRL-and DQN-based approaches showed much better efficiency and robustness (Wang et al [155], Bouktif et al [9]) relative to other baselines.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…One-hundred forty-four fully fledged research articles were finally selected based on the following inclusion criteria: most relevant, most cited, and most recent. For traffic state forecasting, in terms of performances, the GAN-based methods and also hybrid approaches showed better performance on state-of-the-art datasets, i.e., PeMS (Li et al [154], Zhang et al [95]). For intersection signal control, DRL-and DQN-based approaches showed much better efficiency and robustness (Wang et al [155], Bouktif et al [9]) relative to other baselines.…”
Section: Discussionmentioning
confidence: 99%
“…Luo et al [153] combined DL and RL by utilizing the MDP and CNN, which reduced the queue length by 42.5% relative to DQN. Considering knowledge sharing among the agents, Li et al [154] proposed the Knowledge-Sharing Deep Deterministic Policy Gradient (KS-DDPG) algorithm, which showed significant efficiency in controlling large-scale networks and coping with fluctuations in traffic flow. The inability of DRL algorithms to meet the demands of coordination among the agents inspired Wang et al [155] to propose a Cooperative Group-Based Multi-agent reinforcement learning-ATSC (CGB-MATSC) framework that demonstrated a significant reduction of average waiting time by 42.08% relative to FT. Kekuda et al [156] proposed an n-step State, Action, Reward, State, and Action (SARSA) algorithm to increase the implementability in low-cost real-time systems and compared it with LQF; it showed a 5.5% reduction of the average queue length.…”
Section: Neural Network Controllermentioning
confidence: 99%
See 1 more Smart Citation
“…Yao et al proposed a dynamic predictive control framework for traffic signal control in a cross-sectional vehicle infrastructure integration environment [11]. Li et al proposed a multi-agent reinforcement learning method to achieve optimal traffic control by enhancing the cooperation between traffic signals [12]. By introducing the knowledge-sharing enabled communication protocol, each agent can access to the collective representation of the traffic environment collected by all agents.…”
Section: Related Workmentioning
confidence: 99%
“…In [26], the authors investigated a multi-agent algorithm based on Q-learning, taking into account the traffic state at neighboring intersections. In [27], the authors proposed using a knowledge exchange protocol between agents to increase the level of cooperation between agents and achieve an optimal traffic light control strategy. A double Q-learning algorithm for improving the stability of control policy was investigated in [28].…”
Section: Related Workmentioning
confidence: 99%