1994
DOI: 10.1007/bf02830892
|View full text |Cite
|
Sign up to set email alerts
|

Absolutely expedient algorithms for learning Nash equilibria

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

1996
1996
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 9 publications
0
2
0
Order By: Relevance
“…Moreover, the term (R ( ) − R ( −1) ) used in this model makes this algorithm highly robust in tracking a non-stationary job request arrival process. The PDFs are continuously updated by the cloudlets based on their private information and rewards received at every time-slot to learn the pure-strategy NE of the non-cooperative load balancing game and the space complexity of this algorithm is O(( × ) 2 × ), where is the length of memory required for storing the discrete version of ( ) [36].…”
Section: A Distributed Reinforcement Learning Algorithmmentioning
confidence: 99%
“…Moreover, the term (R ( ) − R ( −1) ) used in this model makes this algorithm highly robust in tracking a non-stationary job request arrival process. The PDFs are continuously updated by the cloudlets based on their private information and rewards received at every time-slot to learn the pure-strategy NE of the non-cooperative load balancing game and the space complexity of this algorithm is O(( × ) 2 × ), where is the length of memory required for storing the discrete version of ( ) [36].…”
Section: A Distributed Reinforcement Learning Algorithmmentioning
confidence: 99%
“…) used in this model makes this algorithm highly robust in tracking a non-stationary job request arrival process. The PDFs are continuously updated by the cloudlets based on their private information and rewards received at every time-slot to learn the pure-strategy NE of the non-cooperative load balancing game [40]. Theorem 6.1.…”
mentioning
confidence: 99%