2019
DOI: 10.1109/access.2019.2929120
|View full text |Cite
|
Sign up to set email alerts
|

Target Search Control of AUV in Underwater Environment With Deep Reinforcement Learning

Abstract: The autonomous underwater vehicle (AUV) is widely used to search for unknown targets in the complex underwater environment. Due to the unpredictability of the underwater environment, this paper combines the traditional frontier exploration method with deep reinforcement learning (DRL) to enable the AUV to explore the unknown underwater environment autonomously. In this paper, a grid map of the search environment is built by the grid method. The designed asynchronous advantage actor-critic (A3C) network structu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 51 publications
(22 citation statements)
references
References 34 publications
0
16
0
Order By: Relevance
“…At the same time, DRL and dualstream Q-learning are applied to AUV obstacle avoidance and navigation to further optimize the search path. e simulation results show that the method can effectively control AUV to explore unknown environments [104].…”
Section: Human-inspired Algorithmsmentioning
confidence: 96%
See 1 more Smart Citation
“…At the same time, DRL and dualstream Q-learning are applied to AUV obstacle avoidance and navigation to further optimize the search path. e simulation results show that the method can effectively control AUV to explore unknown environments [104].…”
Section: Human-inspired Algorithmsmentioning
confidence: 96%
“…It can improve the AUV's intelligence level and realize the path planning of the AUV in complex and unknown environments. In recent years, RL methods (such as Q-learning [99,100] and Sarsa [19]) or deep reinforcement learning methods [104] have achieved excellent results in AUV path planning. In 2018, Cao et al used reinforcement learning and Gaussian process regression to solve the path planning with bathymetric aids and modelled the value function as a Gaussian process to minimize the location uncertainty when the AUV reaches the target point [113].…”
Section: Direction C: Intelligent Path Planning Algorithmsmentioning
confidence: 99%
“…The training process uses the DQL algorithm. We have done a preliminary study on the DRL algorithm for path planning, and its specific working process is shown in [39].…”
Section: Path Planning Of Huntingmentioning
confidence: 99%
“…They are represented by dots in different colors. The original position of the target is (39,41,11). It's represented by a purple star.…”
Section: A Target Hunting When Auv Is the Same Velocity As The Targetmentioning
confidence: 99%
See 1 more Smart Citation