2019 IEEE Intelligent Transportation Systems Conference (ITSC) 2019
DOI: 10.1109/itsc.2019.8917533
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Operate a Fleet of Cars

Abstract: In a mobility-on-demand system, travel requests are handled by a fleet of shared vehicles in an on-demand fashion. An important factor that determines the operational efficiency and service level of such a mobility-on-demand system is its operational policy that assigns available vehicles to open passenger requests and relocates idle vehicles. Previously described operational policies are based on control theoretical approaches, most notably on receding horizon control. In this work, we employ reinforcement le… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 18 publications
(13 citation statements)
references
References 27 publications
0
12
0
Order By: Relevance
“…These different changes in waiting times with different parking strategies and distributions can also be explained as follows. While typically in fleet control vehicles are rebalanced to locations of future anticipated demand [33,34,52,53] no rebalancing logic is used on top of the basic global bipartite matching in the present case. Idle and staying vehicles are merely sent to available parking spaces by the parking operating policies.…”
Section: Resultsmentioning
confidence: 99%
“…These different changes in waiting times with different parking strategies and distributions can also be explained as follows. While typically in fleet control vehicles are rebalanced to locations of future anticipated demand [33,34,52,53] no rebalancing logic is used on top of the basic global bipartite matching in the present case. Idle and staying vehicles are merely sent to available parking spaces by the parking operating policies.…”
Section: Resultsmentioning
confidence: 99%
“…2) Cascaded Q-Learning (CQL) [10]: rebalancing policies are learned through tabular Q learning where the current distribution of idle vehicles is chosen as the sole representation of the state space. To achieve a more scalable state-action size, this approach introduces a hierarchical representation of the transportation network, thus learning behavior policies from coarse to finer levels.…”
Section: Methodsmentioning
confidence: 99%
“…As illustrated in Figure 2, we adopt a three-step decisionmaking framework similar to [10] to control an AMoD fleet:…”
Section: A a Three-step Frameworkmentioning
confidence: 99%
See 2 more Smart Citations