2018
DOI: 10.3390/a11020013
|View full text |Cite
|
Sign up to set email alerts
|

muMAB: A Multi-Armed Bandit Model for Wireless Network Selection

Abstract: Multi-armed bandit (MAB) models are a viable approach to describe the problem of best wireless network selection by a multi-Radio Access Technology (multi-RAT) device, with the goal of maximizing the quality perceived by the final user. The classical MAB model does not allow, however, to properly describe the problem of wireless network selection by a multi-RAT device, in which a device typically performs a set of measurements in order to collect information on available networks, before a selection takes plac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
3
1

Relationship

1
9

Authors

Journals

citations
Cited by 23 publications
(16 citation statements)
references
References 23 publications
0
15
0
1
Order By: Relevance
“…Measure-use MAB (muMAB) is proposed in [214] to better adapt MAB models to RAT selection. Classic MABs enable one possible action type in both exploration and exploitation phases, that is, to select an arm and collect the corresponding reward.…”
Section: ) Stateless Mabmentioning
confidence: 99%
“…Measure-use MAB (muMAB) is proposed in [214] to better adapt MAB models to RAT selection. Classic MABs enable one possible action type in both exploration and exploitation phases, that is, to select an arm and collect the corresponding reward.…”
Section: ) Stateless Mabmentioning
confidence: 99%
“…However, the paper does not address the well-known problems of MPTCP such as network middle boxes and TCP modifiers. Boldrini et al [35] tries to answer the following question: Which wireless network can offer the best performance based on the quality observed by end users between the multiple available wireless networks?. To answer this question, the paper laid the foundations based on two main steps: (1) define QoS/QoE parameters; and (2) define the network selection algorithm.…”
Section: Mobile Wireless Video Streamingmentioning
confidence: 99%
“…Contextual bandits are a subset of RL algorithms that are considerably simpler: only one step exists before the outcome is observed. The contextual bandit is an extension of the multiarmed bandit approach [28] wherein the context or state information is considered. Unlike in the multiarmed bandits, the state affects how a reward is associated with each action, and therefore, as the states change, the model needs to learn to adapt its action choice.…”
Section: A Reinforcement Learning-based Rrh Selectionmentioning
confidence: 99%