2020 International Conference on Wireless Communications and Signal Processing (WCSP) 2020
DOI: 10.1109/wcsp49889.2020.9299725
|View full text |Cite
|
Sign up to set email alerts
|

Social Bandit Learning: Strangers Can Help

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…To assess the performance of our social learning algorithm in comparison to alternative methods, we created various scenarios involving learning from non-learners or different types of individual learners. In this section, we compared the performance of our method, SBL-FE, with TUCB [27], OUCB [28] (as social learning algorithms), TS, and UCB (as individual learning methods), using the cumulative regret criteria. We used the same hyperparameters for OUCB and TUCB as stated in their respective papers, and for all subsequent results, we employed the same hyperparameter set.…”
Section: A the Ability Of Social Learning Methods In Different Societiesmentioning
confidence: 99%
See 1 more Smart Citation
“…To assess the performance of our social learning algorithm in comparison to alternative methods, we created various scenarios involving learning from non-learners or different types of individual learners. In this section, we compared the performance of our method, SBL-FE, with TUCB [27], OUCB [28] (as social learning algorithms), TS, and UCB (as individual learning methods), using the cumulative regret criteria. We used the same hyperparameters for OUCB and TUCB as stated in their respective papers, and for all subsequent results, we employed the same hyperparameter set.…”
Section: A the Ability Of Social Learning Methods In Different Societiesmentioning
confidence: 99%
“…Our work is mostly related to [27] and [28], who proposed a social bandit learning algorithm inspired by the Upper Confidence Bound (UCB) learning method to enhance agents' decisions by considering other agents' actions. Both methods are based on the optimism principle about the average of observed policies.…”
Section: Related Workmentioning
confidence: 99%