2013
DOI: 10.1007/s10489-013-0455-3
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical control of traffic signals using Q-learning with tile coding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 57 publications
(34 citation statements)
references
References 17 publications
0
34
0
Order By: Relevance
“…The results demonstrated that the policy gradient actor-critic method consistently outperforms the continuous Q-learning algorithm. Abdoos et al (2014) employed the integration of discrete RL and continuous RL to control 9 traffic signals on a 3 × 3 junction grid. There are two types of agents: bottom level agents and top level agents.…”
Section: Related Workmentioning
confidence: 99%
“…The results demonstrated that the policy gradient actor-critic method consistently outperforms the continuous Q-learning algorithm. Abdoos et al (2014) employed the integration of discrete RL and continuous RL to control 9 traffic signals on a 3 × 3 junction grid. There are two types of agents: bottom level agents and top level agents.…”
Section: Related Workmentioning
confidence: 99%
“…In [13] the authors propose using a set of Q-learning iterations to approach the optimal solution of load balancing. They also mentioned several methods to control the traffic lights and intersections using different techniques of the artificial intelligence, such as fuzzy rules, predefined rule-based systems, and centralized methods.…”
Section: Automobile Applicationsmentioning
confidence: 99%
“…This solution concept is robust to many forms of noise, and does not require expensive recalculations or a complex teamwork protocol. Additionally, we do not impose an explicitly defined organization and communication structure on top of the system as in some work [57], including work with hierarchies [58,59], and instead only rely on the teamwide performance being broadcast to every team member (the team game assumption).…”
Section: Multiagent Systemsmentioning
confidence: 99%