2012 IEEE Congress on Evolutionary Computation 2012
DOI: 10.1109/cec.2012.6256507
|View full text |Cite
|
Sign up to set email alerts
|

A review of inverse reinforcement learning theory and recent advances

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 35 publications
(17 citation statements)
references
References 29 publications
0
17
0
Order By: Relevance
“…This has the advantage that instead of requiring the developer to explicitly specify a reward function, they simply have to demonstrate the intended behaviour. This can be advantageous since in large and complex tasks, defining an adequate reward function to provide optimal agent behaviour can be both difficult and time consuming [130]. IRL approaches have been shown to not only reduce the amount of time required for design and optimisation, but also improve the system performance by creating more robust reward functions.…”
Section: Simultaneous Lateral and Longitudinal Control Systemsmentioning
confidence: 99%
See 1 more Smart Citation
“…This has the advantage that instead of requiring the developer to explicitly specify a reward function, they simply have to demonstrate the intended behaviour. This can be advantageous since in large and complex tasks, defining an adequate reward function to provide optimal agent behaviour can be both difficult and time consuming [130]. IRL approaches have been shown to not only reduce the amount of time required for design and optimisation, but also improve the system performance by creating more robust reward functions.…”
Section: Simultaneous Lateral and Longitudinal Control Systemsmentioning
confidence: 99%
“…Therefore, hand tuning of the derived reward function may be required to ensure safe behaviour. Lastly, the computational burden of IRL methods can be heavy since they often require iteratively solving reinforcement learning problems with each new reward function derived [130]. Nevertheless, in tasks where an adequately accurate reward function cannot be easily defined, IRL approaches can provide an effective solution.…”
Section: Simultaneous Lateral and Longitudinal Control Systemsmentioning
confidence: 99%
“…A main limit in all kinds of IRL above is that the reward functions are considered to be a linear combination of features. The papers [33] and [34] proposed an extended approach by using a limited set of nonlinear rewards. [19] applies Gaussian processes, which is a kind of Non-parametric methods, to cater for potentially complex non-linear feedback functions Although in principle this extends the IRL paradigm to the entire range of non-linear reward functions, the use of kernel machines makes this method easy to require a large number of reward samples to approximate complex reward functions [35].…”
Section: Inverse Reinforcement Learningmentioning
confidence: 99%
“…This approach has a first stage, wherein a human begins controlling the agent. Then, in a second stage, a reward function is derived with Inverse Reinforcement Learning (IRL) (Ng and Russell 2000;Zhifei and Joo 2012) from the collected demonstrations. Finally, this reward function is used in a standard RL process.…”
Section: Background and Related Workmentioning
confidence: 99%