2020
DOI: 10.1109/access.2020.2987861
|View full text |Cite
|
Sign up to set email alerts
|

Energy Efficient Two-Tier Data Dissemination Based on Q-Learning for Wireless Sensor Networks

Abstract: Green communication for different kinds of wireless networks has begun to receive significant research attention recently. Green communication focuses mainly on the issue of improving energy efficiency substantially. A wireless sensor network (WSN) consists of a large number of randomly and widely deployed sensor nodes, and these nodes themselves have the ability to wireless communicate, detect and process data. Sensor nodes can thus detect their surrounding environment, and transmit related data to a sink via… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 21 publications
(12 citation statements)
references
References 26 publications
0
11
0
Order By: Relevance
“…3.1 Q-learning algorithm to find obstacle free paths Q-learning is a reinforcement learning algorithm. One of its working principles is that the agent (the autonomous robot in our research) learns the best policy to adopt for a given scenario based on its interactions with the environment and the rewards gained [35]. According to its current state, a policy in Q-learning is how the agent chooses to behave, react, or make decisions at a given time.…”
Section: Safety Response Mechanism Methodologymentioning
confidence: 99%
See 2 more Smart Citations
“…3.1 Q-learning algorithm to find obstacle free paths Q-learning is a reinforcement learning algorithm. One of its working principles is that the agent (the autonomous robot in our research) learns the best policy to adopt for a given scenario based on its interactions with the environment and the rewards gained [35]. According to its current state, a policy in Q-learning is how the agent chooses to behave, react, or make decisions at a given time.…”
Section: Safety Response Mechanism Methodologymentioning
confidence: 99%
“…The values in the Q table are computed based on a Bellman Equation [8]. Once the Q table is populated after several interactions, the agent uses the information in the Q table to choose the action giving the highest cumulative reward [35,49]. QL can be applied as a dynamic and incremental programming method to find the optimal strategy for a problem in a stepby-step learning mode.…”
Section: Q-learning Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…The RL algorithm utilizes five main elements: the agent, the environment, the reward, the state, and the action. In its learning process, the agent performs several interactions with its environment by making some actions that will cause a change of state in the environment and result in a positive or a negative reward (penalty) [ 35 ]. Over the years, RL has been the subject of many kinds of research in various applications such as chemical reaction [ 36 ], resource management [ 37 ], traffic-light management [ 38 ], autonomous driving [ 39 ], dam management [ 40 ], surgery [ 41 ] and robotics [ 42 ].…”
Section: Short Theory and Background Overviewmentioning
confidence: 99%
“…QL algorithms can be used to iteratively change the MAC protocol parameters by a defined policy to achieve to a low energy state [32]. The TDMA-based adaptive task scheduling [33] method or two-tier data dissemination schemes based on Q-learning (TTDD-QL) [34] are energy efficient for wireless sensor networks (WSN). A cooperative energy-efficient model is presented in the article [35], where clustering, mobile sink deployment and variable sensing collaboratively improve the network lifetime.…”
Section: Introductionmentioning
confidence: 99%