2022
DOI: 10.1109/twc.2021.3126147
|View full text |Cite
|
Sign up to set email alerts
|

Cooperative Multigroup Broadcast 360° Video Delivery Network: A Hierarchical Federated Deep Reinforcement Learning Approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(7 citation statements)
references
References 31 publications
0
7
0
Order By: Relevance
“…Even if viewpoints are known in advance, dynamic network environments such as data traffic load and processing time require adaptive resource management to ensure playback performance. With stochastic decision-making methods, such as reinforcement learning, it is possible to identify the dynamics of user viewpoint movement and determine which tiled videos to deliver to the corresponding VR device (Hu F. et al, 2022). In addition, the portion of tiled videos with different video qualities transmitted in a given time interval can be adjusted according to the viewpoint movement of a user.…”
Section: Content Selectionmentioning
confidence: 99%
See 1 more Smart Citation
“…Even if viewpoints are known in advance, dynamic network environments such as data traffic load and processing time require adaptive resource management to ensure playback performance. With stochastic decision-making methods, such as reinforcement learning, it is possible to identify the dynamics of user viewpoint movement and determine which tiled videos to deliver to the corresponding VR device (Hu F. et al, 2022). In addition, the portion of tiled videos with different video qualities transmitted in a given time interval can be adjusted according to the viewpoint movement of a user.…”
Section: Content Selectionmentioning
confidence: 99%
“…Immersive communication: Open issues and future directions Despite an increasing amount of studies and solutions for supporting XR, haptic communication, and holographic communication, there exist many open issues to address before immersive communications can popularize. To name a few, synchronization of multi-modal communications, user QoE modeling and enhancement, and intelligent network (Hu M. et al, 2022) and machine learning (Hu F. et al, 2022) methods…”
Section: Communication and Networking Solutionsmentioning
confidence: 99%
“…Even if viewpoints are known in advance, dynamic network environments such as data traffic load and processing time require adaptive resource management to ensure playback performance. With stochastic decision-making methods, such as reinforcement learning, it is possible to identify the dynamics of user viewpoint movement and determine which tiled videos to deliver to the corresponding VR device [185]. In addition, the portion of tiled videos with different video qualities transmitted in a given time interval can be adjusted according to the viewpoint movement of a user.…”
Section: A Extended Realitymentioning
confidence: 99%
“…In addition, lower-layer approaches can take into account application-related information for efficient network resource management, e.g., timely adjusting the amount of radio resources allocated to a user in response to the dynamic sensitivity of the user's perception. Since multi-modal perception data in immersive communications can include personal biometric information of individual users, privacy challenges can arise in Balancing the robustness of viewpoint prediction error and video quality by optimization [186] and machine learning [185] methods…”
Section: A Multi-modal Communicationsmentioning
confidence: 99%
“…DQN is a value-based approach, learning an optimal approximated policy of states mapping to actions π(s) = a by parameterizing and estimating state-action value function Q(s, a; θ) where θ denotes the weight matrix of the primary deep neural networks (DNN) [13]. The hDQN framework integrates hierarchical action-value functions operating at different temporal scales using DQN approach and learns optimal approximated policies π 1 (s) = a, a ∈ A B and π 2 (s) = a, a ∈ A N , respectively [14], [15]. Under our hDQN framework, we consider a broad beam (BB) and narrow beam (NB) DQN agents over the same state space S but different action spaces A B and A N , respectively as shown in Figure 3.…”
Section: Hierarchical Dqn-based Beam Alignmentmentioning
confidence: 99%