2022
DOI: 10.1109/tcyb.2021.3111082
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for Solving the Heterogeneous Capacitated Vehicle Routing Problem

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 55 publications
(27 citation statements)
references
References 50 publications
0
26
0
Order By: Relevance
“…The comparison studies investigating properties of deep learning models for specific tasks such as graph embedding and a solution decoding for VRPs are needed. Online DVRP Vera and Abad [59] mCVRP Zhang et al [51] mVRP with soft TW Sykora et al [62] Multi-agent Mapping Problem Lin et al [95] online ride-sharing Falkner and Schmidt-Thieme [60] mCVRPTW Qin et al [115] VRP Silva et al [76] mVRPTW Van Knippenberg et al [38] mCVRP Bogyrbayeva et al [8] mCVRP with charging Bogyrbayeva et al [116] TSP with Drone Chen et al [99] SDDPVD Lin et al [57] mEVRPTW Gutierrez-Rodríguez et al [102] mVRPTW Bono et al [117] mDSCVRPTW, mSCVRPTW Li et al [68] MMHCVRP, MSHCVRP c) Incorporating Uncertainty and Online Routing: The power of learning models lies in their ability to generalize over data distribution. This advantage can be well exploited to incorporate dynamic elements in routing problems, which is not easy to do with the traditional methods [78].…”
Section: The Future Research Directionsmentioning
confidence: 99%
“…The comparison studies investigating properties of deep learning models for specific tasks such as graph embedding and a solution decoding for VRPs are needed. Online DVRP Vera and Abad [59] mCVRP Zhang et al [51] mVRP with soft TW Sykora et al [62] Multi-agent Mapping Problem Lin et al [95] online ride-sharing Falkner and Schmidt-Thieme [60] mCVRPTW Qin et al [115] VRP Silva et al [76] mVRPTW Van Knippenberg et al [38] mCVRP Bogyrbayeva et al [8] mCVRP with charging Bogyrbayeva et al [116] TSP with Drone Chen et al [99] SDDPVD Lin et al [57] mEVRPTW Gutierrez-Rodríguez et al [102] mVRPTW Bono et al [117] mDSCVRPTW, mSCVRPTW Li et al [68] MMHCVRP, MSHCVRP c) Incorporating Uncertainty and Online Routing: The power of learning models lies in their ability to generalize over data distribution. This advantage can be well exploited to incorporate dynamic elements in routing problems, which is not easy to do with the traditional methods [78].…”
Section: The Future Research Directionsmentioning
confidence: 99%
“…The pseudocode of our ITS is presented in Algorithm 1. As for the training algorithm, we use the rollout-based REINFORCE which performs well in training Transformer-style DRL models [41], [42], [49]. The gradient of the loss function is calculated below:…”
Section: Interactive Training Strategymentioning
confidence: 99%
“…Following the data generation in [26], [49], [50], we randomly sample the location of tasks and depot in the uniform distribution U [0, 1]. Two scenarios with four and six UAVs are considered (named U4 and U6, respectively), and each scenario is classified into small-scale situations, medium-scale situations, and large-scale situations according to the problem size.…”
Section: A Experiments Settingsmentioning
confidence: 99%
“…Different from the conventional methods, this line of works aims at automatically searching heuristic policies by using neural networks to learn the underlying patterns in instances, which could be used to discover better policies than hand-crafted ones (Bengio, Lodi, and Prouvost 2021). Towards reducing the gaps to the highly optimized conventional heuristic solvers including Concorde (Applegate et al 2006) and LKH (Helsgaun 2000), a large number of efforts have been performed to invent various deep models to solve the VRP variants, i.e., traveling salesman problem (TSP) and capacitated vehicle routing problem (CVRP) (Khalil et al 2017;Kool, van Hoof, and Welling 2019;Chen and Tian 2019;Hottung and Tierney 2020;Ma et al 2021;Wu et al 2021;Kwon et al 2020;Li et al 2021;Xin et al 2021b).…”
Section: Introductionmentioning
confidence: 99%