2020
DOI: 10.1609/icaps.v30i1.6685
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning Approach to Solve Dynamic Vehicle Routing Problem with Stochastic Customers

Abstract: In real-world urban logistics operations, changes to the routes and tasks occur in response to dynamic events. To ensure customers' demands are met, planners need to make these changes quickly (sometimes instantaneously). This paper proposes the formulation of a dynamic vehicle routing problem with time windows and both known and stochastic customers as a route-based Markov Decision Process. We propose a solution approach that combines Deep Reinforcement Learning (specifically neural networks-based Temporal-Di… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 38 publications
(17 citation statements)
references
References 17 publications
0
12
0
Order By: Relevance
“…The same reward function and rescheduling heuristic are used in the experiment for fairer comparison. Here, we do not include the commonly-used online approaches to solve DVRP such as Multiple Scenario Approach (MSA) (Bent and Van Hentenryck 2004) as baseline because such sampling-based approaches takes longer computation time and may not be operationally suitable (see experiments done in Joe and Lau (2020)).…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The same reward function and rescheduling heuristic are used in the experiment for fairer comparison. Here, we do not include the commonly-used online approaches to solve DVRP such as Multiple Scenario Approach (MSA) (Bent and Van Hentenryck 2004) as baseline because such sampling-based approaches takes longer computation time and may not be operationally suitable (see experiments done in Joe and Lau (2020)).…”
Section: Methodsmentioning
confidence: 99%
“…Recently, many works have emerged that proposed learning-based approaches to solve combinatorial optimization problems, including routing and scheduling problems. There also have been many recent works that addressed the dynamic variants of those problems Joe and Lau 2020;Li et al 2021;Chen, Ulmer, and Thomas 2022). Most of these works adopt a two-stage approach.…”
Section: Reinforcement Learning Approach To Solve Routing and Schedul...mentioning
confidence: 99%
See 1 more Smart Citation
“…Gombolay et al (2018) highlight the value of incorporating human expertise and heuristics in RL performance on VRPs. Joe and Lau (2020) show that combining RL and a meta-heuristic is effective for centralized solving of dynamic VRP problems. In the SOL domain, Irannezhad, Prato, and Hickman (2020) successfully incorporate a multi-agent RL solution into a full-fledged port decision support system and evaluate agent collaboration strategies, showing that a cooperative strategy, rather than one focused on individual reward maximization, results in the highest overall vehicle utilization and lowest travel distance and costs.…”
Section: Reinforcement Learningmentioning
confidence: 98%
“…Because RL has the ability to adapt to dynamic changes in the workload (environment) and handle the non-trivial consequences of chosen policies (actions), it is a good fit for the problems encountered in this paper. We consider cache replacement as a decision-making problem for choosing different replacement policies given the corresponding workload distribution (Joe and Lau 2020). At the same time, to describe the differences between different workloads, we use a neural network (NN) to represent the diverse workload distributions.…”
Section: Deep Reinforcement Learningmentioning
confidence: 99%