2023
DOI: 10.1109/tmm.2021.3132156
|View full text |Cite
|
Sign up to set email alerts
|

Caching in Dynamic Environments: A Near-Optimal Online Learning Approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 45 publications
0
5
0
Order By: Relevance
“…Furthermore, all above works assumed full knowledge of request processes and hence did not incorporate a learning component. A recent line of works considered content caching from an online learning perspective, e.g., [18], [19], and used the performance metric of learning regret or competitive ratio. Works such as [20], [21] used deep RL methods.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Furthermore, all above works assumed full knowledge of request processes and hence did not incorporate a learning component. A recent line of works considered content caching from an online learning perspective, e.g., [18], [19], and used the performance metric of learning regret or competitive ratio. Works such as [20], [21] used deep RL methods.…”
Section: Related Workmentioning
confidence: 99%
“…where β * ∈ R is the minimal long-term average cost of this MDP with parameter W ∈ R, and V (s) is the optimal state value up to an additive constant, which depends on the parameter W. The Q-function can then be defined as [4] Q(s, a) + β * = s−(1−a)W (s) + s p(s |s, a)V (s ), (19) such that V (s) = min a∈{0,1} Q(s, a).…”
Section: A Preliminariesmentioning
confidence: 99%
See 1 more Smart Citation
“…Our machine models are designed to mine the sequential patterns of how each individual user consumes contents and predict for each user which content she will consume the next. Nearly Optimal Cache (NOC) in [51] aims to minimize the dynamic regret, which is the performance gap between an online learning algorithm and the best dynamic policy in hindsight. NOC has provably good worst-case performance for dynamic environments with no prior distribution assumptions, but it potentially degrades the performance when working with friendly request patterns.…”
Section: Related Workmentioning
confidence: 99%
“…(1) LRU-2: evicts content based on the time elapsed since the previous two requests. In PEC, LRU2 is also used to manage the reactive portion; (2) LRU: evicts content based on the time elapsed since the last request; (3) LFU: evicts content based on the request frequency in the whole history; (4) LRB: Learning Relaxed Belady, an online learning approach using the concept of Belady boundary [32] (5) NOC: an online learning based caching algorithm with worst-case performance guarantee [51]; (6) CEC: dynamically selects reactive caching policies using reinforcement learning [54];…”
Section: Comparison With Reactive Cachingmentioning
confidence: 99%