2023
DOI: 10.3390/electronics12183840
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning-Based 2.5D Multi-Objective Path Planning for Ground Vehicles: Considering Distance and Energy Consumption

Xiru Wu,
Shuqiao Huang,
Guoming Huang

Abstract: Due to the vastly different energy consumption between up-slope and down-slope, a path with the shortest length in a complex off-road terrain environment (2.5D map) is not always the path with the least energy consumption. For any energy-sensitive vehicle, realizing a good trade-off between distance and energy consumption in 2.5D path planning is significantly meaningful. In this paper, we propose a deep reinforcement learning-based 2.5D multi-objective path planning method (DMOP). The DMOP can efficiently fin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 34 publications
0
4
0
Order By: Relevance
“…The Q value of the target network is calculated differently from the DQN-PER-TEAE algorithm. Specifically, in the DDQN-PER-TEAE algorithm, the rule for updating the Q value of the target network is: if episode terminates at time step t + 1, then update according to Equation (19), otherwise update according to Equation (20). By decoupling the action selection and the calculation of the target Q value, the DDQN-PER-TEAE algorithm can reduce the overestimation bias in the DQN-PER-TEAE algorithm.…”
Section: ) Ddqn-per-teaementioning
confidence: 99%
See 2 more Smart Citations
“…The Q value of the target network is calculated differently from the DQN-PER-TEAE algorithm. Specifically, in the DDQN-PER-TEAE algorithm, the rule for updating the Q value of the target network is: if episode terminates at time step t + 1, then update according to Equation (19), otherwise update according to Equation (20). By decoupling the action selection and the calculation of the target Q value, the DDQN-PER-TEAE algorithm can reduce the overestimation bias in the DQN-PER-TEAE algorithm.…”
Section: ) Ddqn-per-teaementioning
confidence: 99%
“…The pseudocode of the DDQN-PER-TEAE algorithm is similar to the DQN-PER-TEAE algorithm, except for changes to the network parameters and update formulas. In DDQN-PER-TEAE, the current and target network parameters are represented by θ and θ , and the update formulas in lines 10 and 16 are changed to Equations ( 21), ( 19) and (20), respectively. Compared to the DDQN-PER algorithm, DDQN-PER-TEAE avoids over-estimation and poor model robustness, and allows the agent to explore or exploit actions based on their environmental knowledge.…”
Section: ) Ddqn-per-teaementioning
confidence: 99%
See 1 more Smart Citation
“…These vehicles often navigate complex off-road terrains, where energy consumption poses limitations on their operational range and duration. Researchers have delved into research on energy storage technology, electrical machine control techniques, and internal combustion engine advancements to enhance energy efficiency in this context [18].…”
Section: Introductionmentioning
confidence: 99%