2019
DOI: 10.1016/j.neucom.2019.05.001
|View full text |Cite
|
Sign up to set email alerts
|

TDPP-Net: Achieving three-dimensional path planning via a deep neural network architecture

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 45 publications
(18 citation statements)
references
References 24 publications
(25 reference statements)
0
18
0
Order By: Relevance
“…According to the target Q-values and their current values estimated by the online network, our loss function is then defined as: (13) where α 1 , α 2 , and α 3 are three scaling factors. The first two terms in Equation (13) are aimed at minimizing the differences between the predicted Q-values and their corresponding target values.…”
Section: Training Frameworkmentioning
confidence: 99%
See 2 more Smart Citations
“…According to the target Q-values and their current values estimated by the online network, our loss function is then defined as: (13) where α 1 , α 2 , and α 3 are three scaling factors. The first two terms in Equation (13) are aimed at minimizing the differences between the predicted Q-values and their corresponding target values.…”
Section: Training Frameworkmentioning
confidence: 99%
“…According to the target Q-values and their current values estimated by the online network, our loss function is then defined as: (13) where α 1 , α 2 , and α 3 are three scaling factors. The first two terms in Equation (13) are aimed at minimizing the differences between the predicted Q-values and their corresponding target values. And because the angular and linear velocities are performed concurrently during the interaction, the last term is designed to minimize the difference between the Q-value estimates corresponding to the two commands.…”
Section: Training Frameworkmentioning
confidence: 99%
See 1 more Smart Citation
“…Learning-based planning algorithms have increasingly become more common [23], [24], [25], [26]. In most learningbased path planning algorithms, imitation learning [27] plays a key role [28]. Neural networks have been used to improve the classic algorithms, for instance by adaptively sampling a particular region of a configuration space in sampling-based algorithms [29].…”
Section: Related Workmentioning
confidence: 99%
“…Reinforcement Learning (RL) approaches such as value iteration networks (VINs) [13], [36], learning-fromdemonstration (LfD) [37], guided policy search (GPS) [36], and universal planning networks (UPN) [38] have also been used for path planning. Wu et al present three-dimensional path planning network (TDPP-Net) [28], which is an end-toend network that predicts 3D actions via 2D CNNs. TDPP-Net learns a policy via supervised imitation learning from the Dijkstra's algorithm.…”
Section: Related Workmentioning
confidence: 99%