2020
DOI: 10.1016/j.conengprac.2019.104222
|View full text |Cite
|
Sign up to set email alerts
|

Low-level autonomous control and tracking of quadrotor using reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 68 publications
(32 citation statements)
references
References 17 publications
0
30
0
1
Order By: Relevance
“…The method’s adaptability for multirotor UAVs is demonstrated. By contrast, the trained RL controllers presented in [ 17 , 18 , 21 ] can only be used for a specific multirotor with the same physical structure and parameters. In our method, the policy neural network output can be converted for each actuator unit according to the dynamic model of various geometric characteristics of the vehicle.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…The method’s adaptability for multirotor UAVs is demonstrated. By contrast, the trained RL controllers presented in [ 17 , 18 , 21 ] can only be used for a specific multirotor with the same physical structure and parameters. In our method, the policy neural network output can be converted for each actuator unit according to the dynamic model of various geometric characteristics of the vehicle.…”
Section: Discussionmentioning
confidence: 99%
“…The relation between Equations ( 9 ) and ( 10 ) implies that given a policy , a better policy can be constructed by satisfying the inequality of Equation ( 9 ). Therefore, the proper estimation of in Equation ( 9 ) is the major topic and the approach in this work is based on [ 18 ].…”
Section: Reinforcement Learning Algorithm and Implementationmentioning
confidence: 99%
See 1 more Smart Citation
“…This is one of the most promising areas of machine learning for autonomous control of vehicles. A model-free reinforcement algorithm developed by Chen-Huan Pi et al (2020) has a quad-copter tracing a predetermined path subjected disturbances. The algorithm rewards the system when the trajectory is as efficient at following the predetermined path subsequently minimising the error.…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…For an integrated electrical and heating system with wind energy, Zhang demonstrates PPO can optimize under power, demand, and spot electricity price uncertainties to minimize operational costs [29]. Actor-critic algorithms have also been used in optimizing costs for power systems, lot scheduling, hybrid vehicles, autonomous vehicles, brine injection, and drones [30][31][32][33][34][35][36].…”
Section: Literature Reviewmentioning
confidence: 99%