2021
DOI: 10.1016/j.compag.2021.106350
|View full text |Cite
|
Sign up to set email alerts
|

Collision-free path planning for a guava-harvesting robot based on recurrent deep reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
45
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 87 publications
(49 citation statements)
references
References 15 publications
1
45
0
Order By: Relevance
“…Combining two optimization methods also has been studied [39]. Recently, with the development of deep learning, studies on the path planning using the RL have mainly been proposed [3], [6], [7], [9], [10], [11], [14], [15], [16], [17], [40], [41], [42]. They have supposed the specific scenario and set an environment to apply the agent in the path planning.…”
Section: Path Planningmentioning
confidence: 99%
See 1 more Smart Citation
“…Combining two optimization methods also has been studied [39]. Recently, with the development of deep learning, studies on the path planning using the RL have mainly been proposed [3], [6], [7], [9], [10], [11], [14], [15], [16], [17], [40], [41], [42]. They have supposed the specific scenario and set an environment to apply the agent in the path planning.…”
Section: Path Planningmentioning
confidence: 99%
“…It has been widely used in various fields such as robotics [1], [2], [3], drone [4], [5], [6], [7], [8], [9], military service [10], [11], and self-driving car [12], [13]. Recently, reinforcement learning (RL) has been mainly studied for the path planning [3], [7], [9], [10], [11], [14], [15], [16], [17]. To get an optimal solution, it is essential to give enough reward for an agent to reach the goal and to set up a specific environment.…”
Section: Introductionmentioning
confidence: 99%
“…r 2 is the decay rate, which is a negative constant. The normalized process is usually used in the reward function design [45][46][47], and is calculated by I w /I n . I w is the IAE value obtained by subtracting the command speed and the measured output speed.…”
Section: Reward Function Designmentioning
confidence: 99%
“…The current path planning algorithms mainly include the colony algorithms (Liu et al, 2019;Ye et al, 2020;Zhang et al, 2020;Zhu et al, 2020), PSO (Krell et al, 2019;Wang Y. B. et al, 2019;Liu X. H. et al, 2021;Song et al, 2021), A * algorithms (Xiong et al, 2020;Tang et al, 2021;Tullu et al, 2021), artificial potential field methods Azmi and Ito, 2020;Song et al, 2020;Yao et al, 2020), genetic algorithms (Hao et al, 2020;Li K. R. et al, 2021;Wen et al, 2021), fuzzy control algorithms (Guo et al, 2020;Zhi and Jiang, 2020), fast marching algorithms (Sun et al, 2021;Wang et al, 2021;Xu et al, 2021), and deep reinforcement learning algorithms (Li L. Y. et al, 2021;Lin et al, 2021;Xie et al, 2021). PSO is an evolutionary computation algorithm that can be used to find the optimal solution through collaboration and information sharing between individuals in the group, as in path planning, the optimal solution is to find the shortest path.…”
Section: Introductionmentioning
confidence: 99%