2000 IEEE International Symposium on Information Theory (Cat. No.00CH37060)
DOI: 10.1109/isit.2000.866371
|View full text |Cite
|
Sign up to set email alerts
|

Universal linear least-squares prediction

Abstract: An approach to the problem of linear prediction is discussed that is based on recent developments in the universal coding and computational learning theory literature. This development provides a novel perspective on the adaptive filtering problem, and represents a significant departure from traditional adaptive filtering methodologies. In this context, we demonstrate a sequential algorithm for linear prediction whose accumulated squared prediction error, for every possible sequence, is asymptotically as small… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 34 publications
0
8
0
Order By: Relevance
“…In machine learning literature [1], [2], the area of online learning [3] is heavily investigated in various fields from game theory [4], [5], control theory [6]- [8], decision theory [9], [10] to computational learning theory [11], [12]. Because of the heavily utilized universal prediction perspective [13], it has been considerably applied in data and signal processing [14]- [19], especially in sequential prediction and estimation problems [20]- [23] such as the problem of density estimation and anomaly detection [24]- [28]. Some of its most prominent applications are in multi-agent systems [29]- [31] and specifically in reinforcement learning problems [32]- [42].…”
Section: A Preliminariesmentioning
confidence: 99%
See 1 more Smart Citation
“…In machine learning literature [1], [2], the area of online learning [3] is heavily investigated in various fields from game theory [4], [5], control theory [6]- [8], decision theory [9], [10] to computational learning theory [11], [12]. Because of the heavily utilized universal prediction perspective [13], it has been considerably applied in data and signal processing [14]- [19], especially in sequential prediction and estimation problems [20]- [23] such as the problem of density estimation and anomaly detection [24]- [28]. Some of its most prominent applications are in multi-agent systems [29]- [31] and specifically in reinforcement learning problems [32]- [42].…”
Section: A Preliminariesmentioning
confidence: 99%
“…The multi-armed bandit is widely considered to be the limited feedback version of the well studied prediction with expert advice [15]- [17], [21], [23]. Due to the nature of the problem, only the loss of the selected arm is observed (while others remain hidden).…”
Section: A Preliminariesmentioning
confidence: 99%
“…In these applications of reinforcement learning, we encounter the fundamental dilemma of exploration-exploitation tradeoff, which is most thoroughly studied in the multiarmed bandit problem [8]. The multiarmed bandit problem is generally considered to be the limited feedback version of the wellstudied prediction with expert advice [9]- [13]. It has attracted a significant attention, since the bandit setting can be successfully applied to a wide range of learning applications from recommender systems [14] and dimensionality reduction [15] to probability matching [16].…”
Section: A Preliminariesmentioning
confidence: 99%
“…By summing the probabilities of strategies that suggests the same bandit arm, we construct the probabilities of each bandit arm at time t. The strategies to be used are not specifically selected a priori. Instead, at each time t, all of the strategies s t that compromise the class M t are treated as experts in our online learning problem [11], [13], [21]. These strategies (or experts) are combined according to their weights w s t , which indicates our trust in different strategies, to achieve the performance of the optimal expert.…”
Section: A Brute Force Approachmentioning
confidence: 99%
“…if n > k + 1, and the all-zero vector otherwise. It can be shown using a recursive technique (see e.g., Tsypkin [29], Györfi [15], Singer and Feder [27], and Györfi and Lugosi [18]) that the c n,j can be calculated with small computational complexity. The experts are mixed via an exponential weighting, which is defined the same way as earlier.…”
Section: Generalized Linear Prediction Strategymentioning
confidence: 99%