2023
DOI: 10.1109/tmlcn.2023.3313988
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Reinforcement Learning-Based Resource Scheduler for Massive MIMO Networks

Qing An,
Santiago Segarra,
Chris Dick
et al.

Abstract: The large number of antennas in massive MIMO systems allows the base station to communicate with multiple users at the same time and frequency resource with multiuser beamforming. However, highly correlated user channels could drastically impede the spectral efficiency that multi-user beamforming can achieve. As such, it is critical for the base station to schedule a suitable group of users in each time and frequency resource block to achieve maximum spectral efficiency while adhering to fairness constraints a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 43 publications
0
3
0
Order By: Relevance
“…Considering the complexity of the resource management problem. High-dimensional tasks are generally challenging to deal with for the DRL model due to a phenomenon known as the curse of dimensionality [13]. We propose the DDPG-k algorithm to perform resource allocation because it has better learning in high-dimensional action spaces.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Considering the complexity of the resource management problem. High-dimensional tasks are generally challenging to deal with for the DRL model due to a phenomenon known as the curse of dimensionality [13]. We propose the DDPG-k algorithm to perform resource allocation because it has better learning in high-dimensional action spaces.…”
Section: Related Workmentioning
confidence: 99%
“…However, resource allocation is a discrete decision problem where the combination of the resource allocation cannot exceed the available total value. Inspired by the approach in [13], we use the KNN to discretize DDPG to adapt it to discrete action spaces. The basic idea is to use a continuous-based algorithm to generate an initial continuous action first.…”
Section: Deep Deterministic Policy Gradient Algorithmmentioning
confidence: 99%
See 1 more Smart Citation