2022
DOI: 10.48550/arxiv.2201.05830
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Physical Derivatives: Computing policy gradients by physical forward-propagation

Abstract: Model-free and model-based reinforcement learning are two ends of a spectrum. Learning a good policy without a dynamic model can be prohibitively expensive. Learning the dynamic model of a system can reduce the cost of learning the policy, but it can also introduce bias if it is not accurate. We propose a middle ground where instead of the transition model, the sensitivity of the trajectories with respect to the perturbation of the parameters is learned. This allows us to predict the local behavior of the phys… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 18 publications
(26 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?