2020
DOI: 10.1016/j.ejtl.2020.100008
|View full text |Cite
|
Sign up to set email alerts
|

On modeling stochastic dynamic vehicle routing problems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 83 publications
(21 citation statements)
references
References 72 publications
0
21
0
Order By: Relevance
“…Figure 7 indicates that the convergence of CAV of the RL started from step 2,400 for the actor model and step 4,200 for the A3C algorithm. In each training and testing step, these CAV were calculated using Equation (32). Based on Figure 7, it appears that the agent trained by the A3C algorithm yielded the highest CAV for the performed tasks, because the A3C algorithm had more opportunities to explore the transport environment than the conventional actor-critic algorithm.…”
Section: Experimental Results From the Case Study Without Uncertaintymentioning
confidence: 99%
See 2 more Smart Citations
“…Figure 7 indicates that the convergence of CAV of the RL started from step 2,400 for the actor model and step 4,200 for the A3C algorithm. In each training and testing step, these CAV were calculated using Equation (32). Based on Figure 7, it appears that the agent trained by the A3C algorithm yielded the highest CAV for the performed tasks, because the A3C algorithm had more opportunities to explore the transport environment than the conventional actor-critic algorithm.…”
Section: Experimental Results From the Case Study Without Uncertaintymentioning
confidence: 99%
“…Generally, an RL reward function for VRP uses a positive and negative route cost for the reward and penalizes actions accordingly. This type of reward functions was developed in [29], [31], [32]. However, more aspects are required for a thorough evaluation when considering real-world problems not limited only to routing costs.…”
Section: Vrp Models That Combines ML and Rlmentioning
confidence: 99%
See 1 more Smart Citation
“…The difference between the dynamic vehicle routing problem with stochastic demands (DVRPSD) with VRPSD lies in the different solving perspective, that is, whether the solving process reflects the dynamic adjustment of vehicle route or not [53][54][55]. [56] studied the route planning problem of multi-stage technicians based on experience service time, [57] proposed a RA framework, [58] studied the VRP with real-time traffic information, and [59] proposed a route-based MDP modeling framework.…”
Section: Othersmentioning
confidence: 99%
“…On the other hand, the dynamic stochastic VRP benefits from the problem's prior knowledge. Three strategies, including stochastic modelling [29], sampling method and look-ahead dynamic routing [30] were mainly used in the literature.…”
Section: Related Work and Motivationmentioning
confidence: 99%