2017 15th International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt) 2017
DOI: 10.23919/wiopt.2017.7959912
|View full text |Cite
|
Sign up to set email alerts
|

Deep reinforcement learning-based scheduling for roadside communication networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
39
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 62 publications
(42 citation statements)
references
References 13 publications
0
39
0
1
Order By: Relevance
“…Here, the throughput in the reward function is the received instantaneous throughput, hence the UE with the best channel condition tends to be chosen. The f airness is calculated through Equation (5). α and β are weighting factors for throughput and fairness, respectively.…”
Section: Learning Methodologymentioning
confidence: 99%
“…Here, the throughput in the reward function is the received instantaneous throughput, hence the UE with the best channel condition tends to be chosen. The f airness is calculated through Equation (5). α and β are weighting factors for throughput and fairness, respectively.…”
Section: Learning Methodologymentioning
confidence: 99%
“…Analysis [17], [112], [235]- [266] [73], [97], [187], [267]- [291] Mobility Analysis [227], [292]- [310] User Localization [272], [273], [311]- [315] [111], [316]- [334] Wireless Sensor Networks [335]- [346], [346]- [356] Network Control [186], [293], [357]- [368] [234], [368]- [403] Network Security [185], [345], [404]- [419] [223], [420]- [429], [429]- [436] Signal Processing [378], [380], [437]- [444] [322], [445]- [458] Emerging Applications For each domain, we summarize work broadly in tabular form, providing readers with a general picture of individual topics. Most important works in each domain are discussed in more details in text.…”
Section: App-level Mobile Datamentioning
confidence: 99%
“…In [25], the downlink scheduling has been optimized for battery-charged roadside units in vehicular networks using RL methods to maximize the number of fulfilled service requests during a discharge period, where Q learning is employed to obtain the highest long-term returns. The framework has been further extended in [26], where a deep RL based scheme has been proposed to learn a scheduling policy with high dimensional continuous inputs using end-to-end learning. A distributed user association approach based on RL has been developed in [27] for vehicular networks with heterogeneous BSs.…”
Section: B Related Workmentioning
confidence: 99%