2019
DOI: 10.1609/aaai.v33i01.33013134
|View full text |Cite
|
Sign up to set email alerts
|

State Abstraction as Compression in Apprenticeship Learning

Abstract: State abstraction can give rise to models of environments that are both compressed and useful, thereby enabling efficient sequential decision making. In this work, we offer the first formalism and analysis of the trade-off between compression and performance made in the context of state abstraction for Apprenticeship Learning. We build on Rate-Distortion theory, the classic Blahut-Arimoto algorithm, and the Information Bottleneck method to develop an algorithm for computing state abstractions that approximate … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
64
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
3
1

Relationship

1
9

Authors

Journals

citations
Cited by 47 publications
(68 citation statements)
references
References 37 publications
0
64
0
Order By: Relevance
“…Our recently completed work extends the above state abstraction theory to an information theoretic framework (Abel et al 2019). We draw a parallel between compression, as understood in Information Theory, and state abstraction, as studied in RL, to offer the first formalism and analysis of the trade-off between compression and performance made by state abstraction.…”
Section: Current Workmentioning
confidence: 99%
“…Our recently completed work extends the above state abstraction theory to an information theoretic framework (Abel et al 2019). We draw a parallel between compression, as understood in Information Theory, and state abstraction, as studied in RL, to offer the first formalism and analysis of the trade-off between compression and performance made by state abstraction.…”
Section: Current Workmentioning
confidence: 99%
“…Influence-based abstraction is a form of state abstraction, which has a long tradition in AI planning and learning (e.g., Sacerdoti, 1974;Knoblock, 1993;McCallum, 1993;Dearden & Boutilier, 1997;Hoey, St-Aubin, Hu, & Boutilier, 1999;Givan, Leach, & Dean, 2000;Boutilier, Dearden, & Goldszmidt, 2000;Ravindran & Barto, 2003;Jong & Stone, 2005;Konidaris & Barto, 2009;Kaelbling & Lozano-Perez, 2012;Hostetler, Fern, & Dietterich, 2014;Anand, Noothigattu, Mausam, & Singla, 2016;Bai, Srivastava, & Russell, 2016;Abel, Arumugam, Asadi, Jinnai, Littman, & Wong, 2019) . Other types of abstraction (Mahadevan, 2010) are temporal abstractions, such as options and macro-actions (Sutton, Precup, & Singh, 1999;Theocharous & Kaelbling, 2004;Amato, Konidaris, Kaelbling, & How, 2019;Machado, Bellemare, & Bowling, 2017), and functional abstraction, which tries to identify appropriate basis functions (Keller, Mannor, & Precup, 2006;Parr, Painter-Wakefield, Li, & Littman, 2007;Mahadevan & Maggioni, 2007;Petrik, 2007), including the huge body of recent work on deep RL (Schmidhuber, 1991;Mnih et al, 2015;François-Lavet, Henderson, Islam, Bellemare, & Pineau, 2018).…”
Section: Other Forms Of Abstractionmentioning
confidence: 99%
“…When the cumulative reward function expectations generated by all the strategies are not greater than the cumulative reward function expectations generated by the expert strategy, the reward function of RL will be the reward function learned from the expert data. Apprenticeship learning [30,31] is a type of IRL, which sets the prior basis function as the reward function. This ensures that the optimal strategy obtained from the reward function is near the expert strategy using the given expert data.…”
Section: Inverse Reinforcement Learning (Irl)mentioning
confidence: 99%