2009 Chinese Control and Decision Conference 2009
DOI: 10.1109/ccdc.2009.5194820
|View full text |Cite
|
Sign up to set email alerts
|

RL-based superframe order adaptation algorithm for IEEE 802.15.4 networks

Abstract: In wireless sensor networks, it is an important problem to adjust the work time window in each working/sleeping period to save energy under light network loads and decrease the packet delay under heavy network loads. In this paper, we introduce reinforcement learning method into this problem. We discuss the algorithm design method in a simple IEEE 802.15.4 network, where an RL-based adaptive algorithm is proposed. Simulation results show that this RL-based algorithm can adapt to the change of data flow and mak… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2014
2014
2014
2014

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(6 citation statements)
references
References 9 publications
0
6
0
Order By: Relevance
“…(iii) In comparison to the RL based approaches [18,22,23] for transmission scheduling at the MAC layer, we would like to point out that the algorithms proposed there (a) employ full state representations; (b) consider discrete state-action spaces (except [23] which adapts Qlearning for continuous actions, albeit with a discrete state space); (c) consider an MDP with perfect information, i.e., a setting where the states are fully observable; (d) consider only a discounted setting, which is not amenable for studying steady state system behaviour; (e) are primarily concerned with managing transmission in an energy-efficient manner and not with tracking an intruder with highaccuracy. In other words, the algorithms of [18,22,23] are not applicable in our setting as we consider a partially observable MDP with continuous state-action spaces, and 1 A short version of this paper containing only the average cost setting and algorithms and with no proofs is available in [24]. The current paper includes in addition: (i) algorithms for the discounted cost setting; (ii) a detailed proof of convergence of the average cost algorithm using theory of stochastic recursive inclusions; and (iii) detailed numerical experiments.…”
Section: Related Workmentioning
confidence: 98%
See 4 more Smart Citations
“…(iii) In comparison to the RL based approaches [18,22,23] for transmission scheduling at the MAC layer, we would like to point out that the algorithms proposed there (a) employ full state representations; (b) consider discrete state-action spaces (except [23] which adapts Qlearning for continuous actions, albeit with a discrete state space); (c) consider an MDP with perfect information, i.e., a setting where the states are fully observable; (d) consider only a discounted setting, which is not amenable for studying steady state system behaviour; (e) are primarily concerned with managing transmission in an energy-efficient manner and not with tracking an intruder with highaccuracy. In other words, the algorithms of [18,22,23] are not applicable in our setting as we consider a partially observable MDP with continuous state-action spaces, and 1 A short version of this paper containing only the average cost setting and algorithms and with no proofs is available in [24]. The current paper includes in addition: (i) algorithms for the discounted cost setting; (ii) a detailed proof of convergence of the average cost algorithm using theory of stochastic recursive inclusions; and (iii) detailed numerical experiments.…”
Section: Related Workmentioning
confidence: 98%
“…(iv) Many RL based approaches proposed earlier for sleep scheduling (see [18,22,23,29]) employ full state representations and hence, they are not scalable to larger networks owing to the curse of dimensionality. We employ efficient linear approximators to alleviate this.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations