2019
DOI: 10.1109/twc.2019.2938755
|View full text |Cite
|
Sign up to set email alerts
|

Markov Decision Policies for Dynamic Video Delivery in Wireless Caching Networks

Abstract: This paper proposes a video delivery strategy for dynamic streaming services which maximizes time-average streaming quality under a playback delay constraint in wireless caching networks. The network where popular videos encoded by scalable video coding are already stored in randomly distributed caching nodes is considered under adaptive video streaming concepts, and distance-based interference management is investigated in this paper. In this network model, a streaming user makes delay-constrained decisions d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
22
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
2

Relationship

3
6

Authors

Journals

citations
Cited by 56 publications
(22 citation statements)
references
References 39 publications
(62 reference statements)
0
22
0
Order By: Relevance
“…Angelo furfaro et. al., [32] has simulated the bandwidth flooding attacks handled by the hybrid filter mechanism developed involving differentiated service policies integrated with stopIt technique exploiting network filters on routers. An alerts has been sent by the router to the server when the traffic flow exceeds bandwidth and looks for malicious sources by log monitoring but it is not realistic to large scale topology having more response time thus decrease in performance occurred.…”
Section: Related Workmentioning
confidence: 99%
“…Angelo furfaro et. al., [32] has simulated the bandwidth flooding attacks handled by the hybrid filter mechanism developed involving differentiated service policies integrated with stopIt technique exploiting network filters on routers. An alerts has been sent by the router to the server when the traffic flow exceeds bandwidth and looks for malicious sources by log monitoring but it is not realistic to large scale topology having more response time thus decrease in performance occurred.…”
Section: Related Workmentioning
confidence: 99%
“…The Markov decision process (MDP) can be formed as (S, A, p(s), p(s |s, a), r(s, a, s ), γ) where S and A stand for the sets of states and actions, respectively [40], [41], [42]. Here, γ represents the reflection rate of future rewards compared to the current decision.…”
Section: Background a Markov Decision Processmentioning
confidence: 99%
“…The application of deep neural network overcomes the size of memory that has been the limitation of Q-table in Q-learning; and an efficient function output is possible for large amounts of data. The expressions (5) show that it minimizes a difference between the optimal value and real value (i.e., predicted value). The optimal value is a target value that the agent aims to get.…”
Section: Deep Reinforcement Learningmentioning
confidence: 99%
“…Among them, this paper considers reinforcement learning based approaches because this given problem is for stochastic sequential offloading decision-making. Among a lot of deep reinforcement learning (DRL) methodologies such as Q-learning, Markov decision process (MDP) [5], deep Q-network (DQN), and deep deterministic policy gradient (DDPG) [6,7], this paper designs a sequential offloading decision-making algorithm based on DQN. The reason why this paper considers DQN is that it is the function approximation of Q-learning using deep neural network (DNN) in order to take care of large-scale problem setting.…”
mentioning
confidence: 99%