1994
DOI: 10.1006/game.1994.1057
|View full text |Cite
|
Sign up to set email alerts
|

Computable Strategies for Repeated Prisoner′s Dilemma

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

1996
1996
2015
2015

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(22 citation statements)
references
References 0 publications
0
18
0
Order By: Relevance
“…There has already been some research into the complexity of playing repeated and sequential games. For example, determining whether a particular automaton is a best response is N P-complete (BenPorath, 1990); it is N P-complete to compute a best-response automaton when the automata under consideration are bounded (Papadimitriou, 1992); the problem of whether a given player with imperfect recall can guarantee herself a given payoff using pure strategies is N P-complete (Koller and Megiddo, 1992); and in general, best-responding to an arbitrary strategy can even be noncomputable (Knoblauch, 1994;Nachbar and Zame, 1996). In this section, we present a PSPACE-hardness result on the existence of a pure-strategy equilibrium.…”
Section: Pure-strategy Nash Equilibria In Stochastic (Markov) Gamesmentioning
confidence: 99%
“…There has already been some research into the complexity of playing repeated and sequential games. For example, determining whether a particular automaton is a best response is N P-complete (BenPorath, 1990); it is N P-complete to compute a best-response automaton when the automata under consideration are bounded (Papadimitriou, 1992); the problem of whether a given player with imperfect recall can guarantee herself a given payoff using pure strategies is N P-complete (Koller and Megiddo, 1992); and in general, best-responding to an arbitrary strategy can even be noncomputable (Knoblauch, 1994;Nachbar and Zame, 1996). In this section, we present a PSPACE-hardness result on the existence of a pure-strategy equilibrium.…”
Section: Pure-strategy Nash Equilibria In Stochastic (Markov) Gamesmentioning
confidence: 99%
“…The payoff at every game stage is equally weighted (i.e., there is no discounting for future payoffs). Thus, the payoff can be quantified by the liminf-type asymptotic [26]. The authors then proposed a cooperation strategy via randomized inclination to selfish/greedy play (CRISP) to detect and punish selfish nodes.…”
Section: ) Decentralized Access Methodmentioning
confidence: 99%
“…Any initial play path determines past stage payoffs through (10) and, jointly with the strategy profile , induces a probability distribution of future stage payoffs. If all stage payoffs are equally weighted (there is no discounting of future payoffs), station 's long-term satisfaction, or utility, can be quantified by the following liminf-type asymptotic [15]: (11) where . Since the are bounded, so are their expectations, hence the limit exists and lower bounds the long-term per stage payoff average.…”
Section: B Greedy Selfish and Honest Configurationsmentioning
confidence: 99%
“…If then with probability (tending to one) phase involves a nonzero number of traversals of the left self-loop. For large , the mean of conditioned on the number of stages between a transition to state H and another phase-up can be expressed as (15) where is the number of traversals of the lower left self-loop, , and is a bounded value. Since , and recalling that , we find that for large the right-hand side of (15) asymptotes to , hence tends to as increases.…”
Section: B Pareto Efficiency and Subgame Perfectionmentioning
confidence: 99%
See 1 more Smart Citation