2022
DOI: 10.3390/app12189249
|View full text |Cite
|
Sign up to set email alerts
|

Occupancy Reward-Driven Exploration with Deep Reinforcement Learning for Mobile Robot System

Abstract: This paper investigates the solution to a mobile-robot exploration problem following autonomous driving principles. The exploration task is formulated in this study as a process of building a map while a robot moves in an indoor environment beginning from full uncertainties. The sequence of robot decisions of how to move defines the strategy of the exploration that this paper aims to investigate, applying one of the Deep Reinforcement Learning methods, known as the Deep Deterministic Policy Gradient (DDPG) alg… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 50 publications
0
3
0
Order By: Relevance
“…Related studies proposed by different researchers have been based CME [ 14 ], meta-heuristics [ 37 ], and hybrid methods combining meta-heuristic with deterministic algorithms [ 38 , 39 , 40 ]. Some studies focused solely on static sensor coverage faults and robot movements in uncharted surroundings [ 41 , 42 , 43 ] as well as exploration with deep reinforcement learning for mobile robots [ 44 ]; however, they coincide in meaning but the goal of both studies is to create a finite map. The rest of the paper is organized as follows:…”
Section: Related Workmentioning
confidence: 99%
“…Related studies proposed by different researchers have been based CME [ 14 ], meta-heuristics [ 37 ], and hybrid methods combining meta-heuristic with deterministic algorithms [ 38 , 39 , 40 ]. Some studies focused solely on static sensor coverage faults and robot movements in uncharted surroundings [ 41 , 42 , 43 ] as well as exploration with deep reinforcement learning for mobile robots [ 44 ]; however, they coincide in meaning but the goal of both studies is to create a finite map. The rest of the paper is organized as follows:…”
Section: Related Workmentioning
confidence: 99%
“…Hyperparameter tuning was critical to the success of this method [8]. A different approach, called Occupancy-Reward-Driven Exploration [9], has been applied in robotics to explore uncharted territories within the state space. In this technique, an occupancy map is utilized to acquire information about the environment through sensors such as a laser sensor.…”
Section: Introductionmentioning
confidence: 99%
“…The robot's reward is then determined by the number of new segments discovered within the occupancy map at each time step. This approach can also improve the robot's power efficiency [9].…”
Section: Introductionmentioning
confidence: 99%