The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2021
DOI: 10.1186/s13638-021-01939-x
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic handoff policy for RAN slicing by exploiting deep reinforcement learning

Abstract: It has been widely acknowledged that network slicing is a key architectural technology to accommodate diversified services for the next generation network (5G). By partitioning the underlying network into multiple dedicated logical networks, 5G can support a variety of extreme business service needs. As network slicing is implemented in radio access networks (RAN), user handoff becomes much more complicated than that in traditional mobile networks. As both physical resource constraints of base stations and log… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(2 citation statements)
references
References 14 publications
0
2
0
Order By: Relevance
“…The , per slice is dependent on and values, where their values can be determined either by with initial values or valued from the previous Q-learning algorithm iteration (line 3). In each state, an action is performed based on the ε-greedy policy [40] that represents both the exploration (random action selection) and the exploitation (action with the maximum Q-value) trade-off (lines 4,5). is then calculated according to (20)…”
Section: Rl-based Resource Slicing (Q-learning) Algorithmmentioning
confidence: 99%
“…The , per slice is dependent on and values, where their values can be determined either by with initial values or valued from the previous Q-learning algorithm iteration (line 3). In each state, an action is performed based on the ε-greedy policy [40] that represents both the exploration (random action selection) and the exploitation (action with the maximum Q-value) trade-off (lines 4,5). is then calculated according to (20)…”
Section: Rl-based Resource Slicing (Q-learning) Algorithmmentioning
confidence: 99%
“…Simulation results prove the efficiency of the framework as compared to reactive schemes in reducing unnecessary handovers. User handoff in 5G RAN network slicing has been addressed in [70]. The authors devised an intelligent handoff policy that considers two main constraints: physical resources of base stations and logical connection of network slices.…”
Section: ) Literature Reviewmentioning
confidence: 99%