2019 IEEE International Conference on Robotics and Biomimetics (ROBIO) 2019
DOI: 10.1109/robio49542.2019.8961639
|View full text |Cite
|
Sign up to set email alerts
|

Obstacle Avoidance Using Stereo Vision and Deep Reinforcement Learning in an Animal-like Robot

Abstract: This is a repository copy of Obstacle avoidance using stereo vision and deep reinforcement learning in an animal-like robot.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 16 publications
0
4
0
Order By: Relevance
“…Similarly, a robot may have a separate local navigation module that would generate a course diversion to avoid hitting objects. We have previously demonstrated how this local navigation could be acquired through RL [35,48]. More generally, control architectures in both animals and robots are likely to distinguish between the global and local navigation problems and employ separate mechanisms for each [44].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Similarly, a robot may have a separate local navigation module that would generate a course diversion to avoid hitting objects. We have previously demonstrated how this local navigation could be acquired through RL [35,48]. More generally, control architectures in both animals and robots are likely to distinguish between the global and local navigation problems and employ separate mechanisms for each [44].…”
Section: Discussionmentioning
confidence: 99%
“…The MiRo robot is a commercially available biomimetic robot developed by Consequential Robotics Ltd in partnership with the University of Sheffield. MiRo's physical design and control system architecture find their inspiration in biology, psychology and neuroscience [39], making it a valuable platform for embedded testing of brain-inspired models of perception, memory and learning [35]. For mobility, the robot is differentially driven, whilst we use its front-facing sonar to detect approaching walls and objects for sensing.…”
Section: Miro Robot and The Testing Environmentmentioning
confidence: 99%
“…By employing reinforcement learning, particularly the Dyna-Q approach, quadcopters can enhance their decision-making and adapt their flight trajectories. This combination of strategic path planning and adaptive obstacle avoidance, aided by advanced machine learning, allows quadcopters to optimize their operations, prevent collisions, and maintain stability while dynamically adjusting to their surroundings and achieving mission objectives [27][28][29][30][31].…”
Section: Enhancing Quadcopter Trajectory Tracking Through Dyna-q Lear...mentioning
confidence: 99%
“…Dyna-Q learning combines both learning the model and Q-learning to optimize the learning process effectively. In reinforcement learning, the Markov Decision Process (MDP) is used to model the interactions between an agent and the environment, helping the agent maximize cumulative rewards in uncertain environments [12,27,29,32,33]. MDPs aim to determine policies that guide the agent's actions:…”
Section: Agent Environmentmentioning
confidence: 99%