Proceedings of the 8th ACM Conference on Recommender Systems 2014
DOI: 10.1145/2645710.2645732
|View full text |Cite
|
Sign up to set email alerts
|

Ensemble contextual bandits for personalized recommendation

Abstract: The cold-start problem has attracted extensive attention among various online services that provide personalized recommendation. Many online vendors employ contextual bandit strategies to tackle the so-called exploration/exploitation dilemma rooted from the cold-start problem. However, due to high-dimensional user/item features and the underlying characteristics of bandit policies, it is often difficult for service providers to obtain and deploy an appropriate algorithm to achieve acceptable and robust economi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
46
1

Year Published

2015
2015
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 69 publications
(47 citation statements)
references
References 22 publications
0
46
1
Order By: Relevance
“…Tang et al [225] propose a context-aware recommender system, implemented as a contextual multi-armed bandits problem. Although the authors report extensive offline evaluation (log based and simulation based) with acceptable CTR, no comparison is made from a cold-start problem standpoint.…”
Section: Cold Start Problemmentioning
confidence: 99%
“…Tang et al [225] propose a context-aware recommender system, implemented as a contextual multi-armed bandits problem. Although the authors report extensive offline evaluation (log based and simulation based) with acceptable CTR, no comparison is made from a cold-start problem standpoint.…”
Section: Cold Start Problemmentioning
confidence: 99%
“…The resulting cumulative reward is considered "the unique good metric to evaluate the recommendation algorithm" [74]. It is common to extend this setting to a contextual bandit problem [117,37,192] [114]. This approach is known as ranked bandits.…”
Section: Multi-arm Banditsmentioning
confidence: 99%
“…This type of MAB is called multiple-play bandit and has been studied in [109,123]. Another emerging approach is based on ensemble learning, where the bandit algorithm decides which recommendation model to choose for filling the slots in the top-N ranking via exploring the potential of untested models and exploiting the predictive power of the already tested ones [192,62,31].…”
Section: Multi-arm Banditsmentioning
confidence: 99%
See 2 more Smart Citations