2012 IEEE International Conference on Pervasive Computing and Communications Workshops 2012
DOI: 10.1109/percomw.2012.6197639
|View full text |Cite
|
Sign up to set email alerts
|

Resource coordination in wireless sensor networks by cooperative reinforcement learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
30
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 20 publications
(30 citation statements)
references
References 7 publications
0
30
0
Order By: Relevance
“…Khan and Rinner [8] apply reinforcement learning (RL) for online task scheduling. They use cooperative reinforcement learning for task scheduling.…”
Section: Reinforcement Learning-based Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Khan and Rinner [8] apply reinforcement learning (RL) for online task scheduling. They use cooperative reinforcement learning for task scheduling.…”
Section: Reinforcement Learning-based Methodsmentioning
confidence: 99%
“…Khan and Rinner [8] propose cooperative Q learning (RL) where every agent needs to maintain a Q matrix for the value functions like independent Q learning. Initially, all entries of the Q matrix are 0 and the nodes or agents may be in any state.…”
Section: Rlmentioning
confidence: 99%
See 1 more Smart Citation
“…In [3], the problem of finding an efficient sleep-wake policy for the sensors while maintaining good tracking accuracy by solving an MDP has been studied. In [20], the authors propose a Q-learning based algorithm for sleep scheduling. In [14,15], the authors propose a POMDP model for sleepscheduling in an object tracking application and propose several algorithms based on traditional dynamic programming approaches to solve this problem.…”
Section: Related Workmentioning
confidence: 99%
“…Update θ n using (20) Slow timescale θ n+1 g(s n , a n ) Fig. 3 Overall flow of the TQSA-A algorithm from the perturbed simulation, the gradient of the approximate Q-value function Qðs; aÞ % h T r s;a is estimated as:…”
Section: Fast Timescalementioning
confidence: 99%