2012
DOI: 10.1007/978-3-642-35506-6_35
|View full text |Cite
|
Sign up to set email alerts
|

On Ensemble Techniques for AIXI Approximation

Abstract: One of the key challenges in AIXI approximation is model class approximation -i.e. how to meaningfully approximate Solomonoff Induction without requiring an infeasible amount of computation? This paper advocates a bottom-up approach to this problem, by describing a number of principled ensemble techniques for approximate AIXI agents. Each technique works by efficiently combining a set of existing environment models into a single, more powerful model. These techniques have the potential to play an important rol… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2013
2013
2019
2019

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 12 publications
0
3
0
Order By: Relevance
“…The creation of such an algorithm as a single piece of code is theoretically sound but at best intractable [2]. Various methods have been attempted to deal with the incomputabilityintractability question, to list a few [3], [4]. AGINAO builds its cognitive engine as a hierarchy of interconnected datastructures, named concepts, each with a built-in piece of executable code, named codelet.…”
Section: A the Aginao Self-programming Enginementioning
confidence: 99%
“…The creation of such an algorithm as a single piece of code is theoretically sound but at best intractable [2]. Various methods have been attempted to deal with the incomputabilityintractability question, to list a few [3], [4]. AGINAO builds its cognitive engine as a hierarchy of interconnected datastructures, named concepts, each with a built-in piece of executable code, named codelet.…”
Section: A the Aginao Self-programming Enginementioning
confidence: 99%
“…Even worse, the dependence on C is not an artifact of the analysis, but rather a failing of the EG algorithm, which becomes unstable when experts transition from predicting badly to predicting well. Another algorithm with near-linear running time is online gradient descent (OGD) by Zinkevich [2003] (and applied to this setting by Veness et al [2012a]), which runs in O(N log(N )) time using the fast simplex projection by Duchi et al [2008]. The regret of this algorithm also depends on the size of the maximum gradient of the loss, however, which leads to bound of the same order as Eq.…”
Section: Introductionmentioning
confidence: 99%
“…methods which: (p) make probabilistic predictions; (o) are strongly online; (w) work well in practice; (e) are efficient; (r) and have well understood regret/loss/redundancy properties. Methods satisfying these properties can be combined in a principled fashion using techniques such as those discussed by [VSH12,Mat13], giving rise to ensembles with clearly interpretable predictive capabilities.…”
Section: Introductionmentioning
confidence: 99%