We consider a class of restless multi-armed bandit problems (RMBP) that arises in dynamic multichannel access, user/server scheduling, and optimal activation in multi-agent systems. For this class of RMBP, we establish the indexability and obtain Whittle's index in closed-form for both discounted and average reward criteria. These results lead to a direct implementation of Whittle's index policy with remarkably low complexity. When these Markov chains are stochastically identical, we show that Whittle's index policy is optimal under certain conditions. Furthermore, it has a semi-universal structure that obviates the need to know the Markov transition probabilities. The optimality and the semi-universal structure result from the equivalency between Whittle's index policy and the myopic policy established in this work. For non-identical channels, we develop efficient algorithms for computing a performance upper bound given by Lagrangian relaxation. The tightness of the upper bound and the near-optimal performance of Whittle's index policy are illustrated with simulation examples.
Index TermsOpportunistic access, dynamic channel selection, restless multi-armed bandit, Whittle's index, indexability, myopic policy.Proof: The upper bound of J is obtained from the upper bound of the optimal performance for generally non-identical channels as given in (43). The lower bound of J w is obtained from the structure of Whittle's index policy. See Appendix H for the complete proof.Corollary 2: Let η = Jw J be the approximation factor defined as the ratio of the performance by Whittle's index policy to the optimal performance. We have