2019
DOI: 10.3233/jifs-172180
|View full text |Cite
|
Sign up to set email alerts
|

Intelligent controller for passivity-based biped robot using deep Q network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 28 publications
0
6
0
Order By: Relevance
“…30 Besides, adaptive model-free controllers using RL were proposed to stabilize PDW-based robots, improving robustness to disturbance and time varying environments, while maintain PDW’s merits of human-like motion and high energy efficiency. 31,32 So this paper expanded out previous work to improve the PDW-based robot’s versatility by stabilizing its chaotic behaviors.…”
Section: Introductionmentioning
confidence: 94%
“…30 Besides, adaptive model-free controllers using RL were proposed to stabilize PDW-based robots, improving robustness to disturbance and time varying environments, while maintain PDW’s merits of human-like motion and high energy efficiency. 31,32 So this paper expanded out previous work to improve the PDW-based robot’s versatility by stabilizing its chaotic behaviors.…”
Section: Introductionmentioning
confidence: 94%
“…Humanoid robots are high-dimensional non-smooth systems with many physical constraints. Nowadays, an increasing number of humanoid robot control algorithms are being developed adopting DRL in research on, for example, robot balance [3][4][5], the dynamic performance after the training of a legged robot [6] and the passive dynamic walking robot [7]. Vuga et al [8] used motion capture with a model-free reinforcement learning algorithm to reduce differences in humanoid robots learning human actions.…”
Section: Biped Robot Controlled By Drlmentioning
confidence: 99%
“…Figure 4 shows the basic structure of the AUV hunting path planning implemented by the DRL method. The AUV uses DRL and a dual-stream Q-network to avoid obstacles [37][38]. By setting a negative reward for collision, we pursue a strategy to maximize the total reward.…”
Section: Path Planning Of Huntingmentioning
confidence: 99%
“…A target speed of 2 grids/second. The AUVs complete the hunting at the target marching point (32,38,37). Under the same conditions, the proposed algorithm uses five AUVs to successfully close the target at (28,25,22), and the hunting distance is shortened by 48%, as shown in Figure 9(b).…”
Section: Figure 8 the Structure Diagram Of Bio-inspired Neural Networkmentioning
confidence: 99%