2010 Ninth International Conference on Machine Learning and Applications 2010
DOI: 10.1109/icmla.2010.65
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Agent Inverse Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
39
0

Year Published

2012
2012
2023
2023

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 62 publications
(41 citation statements)
references
References 9 publications
0
39
0
Order By: Relevance
“…Moreover, the demonstrations are performed by multiple experts. Contrary to [Natarajan et al, 2010], the experts' policies are not independent from each other but take other experts into account. The methods developed in [Bogert and Doshi, 2014, Bogert et al, 2016 are based on maximum entropy IRL .…”
Section: Irl From Partially Observable Demonstrationsmentioning
confidence: 99%
“…Moreover, the demonstrations are performed by multiple experts. Contrary to [Natarajan et al, 2010], the experts' policies are not independent from each other but take other experts into account. The methods developed in [Bogert and Doshi, 2014, Bogert et al, 2016 are based on maximum entropy IRL .…”
Section: Irl From Partially Observable Demonstrationsmentioning
confidence: 99%
“…Active learning [Settles, 2012] relies on the fact that if an algorithm can only solicit labels of a limited number of examples, then it should choose them judiciously since not all examples provide the same amount of information. Active learning has a long history of being successfully employed with a variety of classifiers such as logistic regression [Lewis and Gale, 1994;Lewis and Catlett, 1994], support vector machines [Tong and Koller, 2001b], Bayesian network learning [Tong and Koller, 2000;2001a] and in sequential decision making tasks such as imitation learning [Judah et al, 2014] and inverse reinforcement learning [Odom and Natarajan, 2016].…”
Section: Related Workmentioning
confidence: 99%
“…The problem of choosing an example to obtain its class label has been addressed as active learning [Settles, 2012]. There have been several extensions of active learning that included presenting a set of features [Raghavan et al, 2006;Druck et al, 2009], or getting labels over clusters [Hofmann and Buhmann, 1998], or preferences [Odom and Natarajan, 2016] or in sequential decision making [Lopes et al, 2009], to name a few.…”
Section: Introductionmentioning
confidence: 99%
“…Abbeel et al [11] reported the approach of "Apprenticeship learning" in which the optimal policy is acquired in the process of presuming a reward function. Natarajan et al [12] presumed multiple reward functions in a multiagent environment, and proposed an approach for controlling the behaviors from a global perspective.…”
Section: Inverse Reinforcement Learningmentioning
confidence: 99%
“…Various approaches have been proposed [10], [11], [12]. Ng et al [10] reported an approach for estimating a reward function using linear programming for an environment with finite state space, and the Monte Carlo method for an environment with an infinite state space.…”
Section: Inverse Reinforcement Learningmentioning
confidence: 99%