2021
DOI: 10.1007/978-981-16-6320-8_68
|View full text |Cite
|
Sign up to set email alerts
|

Robot Path Planning via Deep Reinforcement Learning with Improved Reward Function

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 8 publications
0
3
0
Order By: Relevance
“…One crucial aspect of DRL algorithms is the reward function, which fundamentally shapes the agent's learning strategy and the direction for network optimization. Crafting an ideal explicit reward function to meet long-term goals is a formidable task, chiefly because mapping relationships from complex state spaces to reward values can be nonlinear, making the manual description of the relationships between reward components highly challenging [33]. Initial research focused predominantly on single-objective optimization, primarily centered on position control [34,35], simplifying the reward function as follows:…”
Section: Cascaded Fuzzy Reward System (Cfrs)mentioning
confidence: 99%
“…One crucial aspect of DRL algorithms is the reward function, which fundamentally shapes the agent's learning strategy and the direction for network optimization. Crafting an ideal explicit reward function to meet long-term goals is a formidable task, chiefly because mapping relationships from complex state spaces to reward values can be nonlinear, making the manual description of the relationships between reward components highly challenging [33]. Initial research focused predominantly on single-objective optimization, primarily centered on position control [34,35], simplifying the reward function as follows:…”
Section: Cascaded Fuzzy Reward System (Cfrs)mentioning
confidence: 99%
“…Examples of the RL-based algorithms utilized in motion planning include the twin delayed deep deterministic policy gradient (TD3) [107] and Exploitation of Abstract Symmetry of Environments (EASE). The latter relies on locally adopting spatial symmetry abstractions obtained from naïvely trained agents [108].…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…As aforementioned, training RL models is challenging in terms of convergence and robustness. The reasons for this include ambiguities in the relationship between the Cartesian and joint spaces, continuous workspaces, and redundant DoF, which result in unnecessary explorations [107]. This could be alleviated using a NN to produce an initial policy for guiding the training of the RL framework [105].…”
Section: Reinforcement Learningmentioning
confidence: 99%