2021 60th IEEE Conference on Decision and Control (CDC) 2021
DOI: 10.1109/cdc45484.2021.9683135
|View full text |Cite
|
Sign up to set email alerts
|

Graph Neural Network Reinforcement Learning for Autonomous Mobility-on-Demand Systems

Abstract: Optimization problems over dynamic networks have been extensively studied and widely used in the past decades to formulate numerous real-world problems. However, (1) traditional optimization-based approaches do not scale to large networks, and (2) the design of good heuristics or approximation algorithms often requires significant manual trial-and-error. In this work, we argue that data-driven strategies can automate this process and learn efficient algorithms without compromising optimality. To do so, we pres… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 34 publications
(11 citation statements)
references
References 64 publications
0
11
0
Order By: Relevance
“…A Three-Step Framework. As in [10,15], we formulate the AMoD control problem as a three-step decision-making framework, whereby the control of an AMoD fleet follows three steps: matching, RL for rebalancing, and post-processing. This three-step framework, as shown in the rest of this section, has the advantage of reducing the action space from 𝑁 2 𝑣 to 𝑁 𝑣 , since the learned policy defines an action at each node as opposed to along each OD pair (as in the majority of literature).…”
Section: Control Of Single-city Amod Systemsmentioning
confidence: 99%
See 3 more Smart Citations
“…A Three-Step Framework. As in [10,15], we formulate the AMoD control problem as a three-step decision-making framework, whereby the control of an AMoD fleet follows three steps: matching, RL for rebalancing, and post-processing. This three-step framework, as shown in the rest of this section, has the advantage of reducing the action space from 𝑁 2 𝑣 to 𝑁 𝑣 , since the learned policy defines an action at each node as opposed to along each OD pair (as in the majority of literature).…”
Section: Control Of Single-city Amod Systemsmentioning
confidence: 99%
“…Related work. Within existing literature, AMoD systems can be coordinated with simple rule-based heuristics [6,7], model predictive control (MPC) approaches [8,9], and RL approaches [10][11][12][13][14][15]. Interested readers can refer to [16] for a comprehensive survey of the analysis and control of AMoD systems and [17] for RL approaches for ridesharing systems.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Electric vehicles (EVs) are being adopted worldwide for environmental and economical benefits [4], and AMoD systems embrace this trend without exception. However, the trips sporadically appear, and the origins and destinations are asymmetrically distributed and hard to predict in AMoD systems [3], [5]. Such spatial-temporal nature of urban mobility increases the management difficulty of a large-scale vehicle fleet and makes the system sensitive to disturbances [5]- [7].…”
Section: Introductionmentioning
confidence: 99%