2023
DOI: 10.1109/access.2023.3255007
|View full text |Cite
|
Sign up to set email alerts
|

Predator-Prey Reward Based Q-Learning Coverage Path Planning for Mobile Robot

Abstract: Coverage Path Planning (CPP in short) is a basic problem for mobile robot when facing a variety of applications. Q-Learning based coverage path planning algorithms are beginning to be explored recently. To overcome the problem of traditional Q-Learning of easily falling into local optimum, in this paper, the new-type reward functions originating from Predator-Prey model are introduced into traditional Q-Learning based CPP solution, which introduces a comprehensive reward function that incorporates three reward… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 31 publications
0
3
0
Order By: Relevance
“…It seamlessly combines these two essential components, where exploration generates a map used by the coverage path planning algorithm. In contrast, several other approaches primarily focus on either exploration or coverage separately, and their explicit integration may not be as robust or evident in those cases [5], [6], [30], [33], [35]. Another standout feature of the proposed algorithm is its commitment to power efficiency.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…It seamlessly combines these two essential components, where exploration generates a map used by the coverage path planning algorithm. In contrast, several other approaches primarily focus on either exploration or coverage separately, and their explicit integration may not be as robust or evident in those cases [5], [6], [30], [33], [35]. Another standout feature of the proposed algorithm is its commitment to power efficiency.…”
Section: Resultsmentioning
confidence: 99%
“…article addresses path planning for multiple UAVs to achieve sweep coverage, especially focusing on forest fire early warning and monitoring. A Predator-Prey reward-based Q-Learning CPP, overcoming local optima challenges is studied in [35] and [36] introduces a visibility-based path planning (VPP) heuristic for optimizing visibility during UAV flights.…”
Section: Motivationmentioning
confidence: 99%
“…Hao et al [49] proposed a dynamic fast Q-learning (DFQL) algorithm for the path planning problem of USV in some known marine environments, which combines Q-learning with artificial potential field (APF) to initialize the Q table and provides USV with prior knowledge from the environment. Zhang et al [50] in order to overcome the problem that traditional Q-learning is prone to local optimization in coverage path planning, a new reward function derived from the predator-prey model is introduced into the traditional Q-learning-based CPP solution.…”
Section: Related Workmentioning
confidence: 99%
“…Deep Q-learning (DQN) is an algorithm improved by Q-Learning [20][21][22]. The expression of the traditional reinforcement Q-learning algorithm is shown in the following formula:…”
Section: Dqnmentioning
confidence: 99%