2022
DOI: 10.3390/pr10071307
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for Integrated Non-Linear Control of Autonomous UAVs

Abstract: In this research, an intelligent control architecture for an experimental Unmanned Aerial Vehicle (UAV) bearing unconventional inverted V-tail design, is presented. To handle UAV’s inherent control complexities, while keeping them computationally acceptable, a variant of distinct Deep Reinforcement Learning (DRL) algorithm, namely Deep Deterministic Policy Gradient (DDPG) is proposed. Conventional DDPG algorithm after being modified in its learning architecture becomes capable of intelligently handling the con… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 23 publications
(9 citation statements)
references
References 54 publications
0
4
0
Order By: Relevance
“…Meanwhile, the environment model reflects the ambient behavior that aids the algorithm's efficiency by enabling it to comprehend its environment. RL algorithms may use agents to explore their environment [31], [148]. When RL is paired with deep learning's superior comprehending capabilities, RL is more efficient at generating decisions than humans in practically unlimited state space [82].…”
Section: Automation and Artificial Intelligencementioning
confidence: 99%
“…Meanwhile, the environment model reflects the ambient behavior that aids the algorithm's efficiency by enabling it to comprehend its environment. RL algorithms may use agents to explore their environment [31], [148]. When RL is paired with deep learning's superior comprehending capabilities, RL is more efficient at generating decisions than humans in practically unlimited state space [82].…”
Section: Automation and Artificial Intelligencementioning
confidence: 99%
“…This method is very efficient when working with large problems involving a lot of data or parameters for which the algorithm is concerned. Adam has advantages over other optimizers, namely adaptive gradient, and root mean square propagation (RMSProp) [37], [38]. In determining the weight value if RMSProp uses the first value in the gradient, on the other hand, Adam uses the second value from the gradient.…”
Section: Adam Optimizermentioning
confidence: 99%
“…UAV trajectory (Din et al, 2022) [80] Techniques: (DDPG) algorithm, modified for learning architecture to handle continuous state and control space domains Methodology: Designing an intelligent control architecture for an experimental Unmanned Aerial Vehicle (UAV) (DDPG) algorithm…”
Section: Uav-bs Systemsmentioning
confidence: 99%