Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining 2018
DOI: 10.1145/3219819.3219824
|View full text |Cite
|
Sign up to set email alerts
|

Large-Scale Order Dispatch in On-Demand Ride-Hailing Platforms

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
216
0
2

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 306 publications
(219 citation statements)
references
References 23 publications
1
216
0
2
Order By: Relevance
“…at is to say, the principle of matching orders based on distance is equivalent to random matching in our environment se ing. • MDP: Proposed by Xu et al [37], a planning and learning method based on decentralized multi-agent deep reinforcement learning and centralized combinatorial optimization.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…at is to say, the principle of matching orders based on distance is equivalent to random matching in our environment se ing. • MDP: Proposed by Xu et al [37], a planning and learning method based on decentralized multi-agent deep reinforcement learning and centralized combinatorial optimization.…”
Section: Methodsmentioning
confidence: 99%
“…But considering the current operational ride-sharing scenarios, it is hard to perform eet management for it is impossible to force drivers to designated regions. e mentioned MARL method [37] is an independent MARL method, which ignores the interactions between agents. However, it is a consensus to consider that the agent interactions have a positive impact on making optimal decisions.…”
Section: Related Workmentioning
confidence: 99%
“…One example of a heuristic solution to the order dispatching problem is the myopic pickup distance minimization (MPDM), which ignores temporal dynamics and always assigns the closest available driver to a requesting order [12]. In Local Policy Improvement [13], handcrafted heuristics are combined with a machine learning method by summarizing supply and demand patterns into a table and then using the learned patterns to account for future gains in the real-time planner of the dispatching solution. A fully machine learning based approach to the dispatching problem is presented in [14], where deep Q-learning is used for learning dispatching strategies from the perspective of a single driver.…”
Section: Related Workmentioning
confidence: 99%
“…One alternative is to model each available vehicle as an agent [21, Forthcoming Order Figure 1: Ride-hailing task in thermodynamics view. 37,39]. However, such setting needs to maintain thousands of agents interacting with the environment, which brings a huge computational cost.…”
Section: Introductionmentioning
confidence: 99%
“…Greedily matching vehicles with long-distance orders can receive high immediate gain at a single order dispatching stage, but it would harm order response rate (ORR) and future revenue especially during rush hour because of its long drive time and unpopular destination. Recent attempts [21,37,39] deployed RL to combine instant order reward from online planning with future state-value as the final matching value. However, the coordination between different regions is still far from optimal.…”
Section: Introductionmentioning
confidence: 99%