2021
DOI: 10.1109/jiot.2020.3005598
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic Charging Scheme Problem With Actor–Critic Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
22
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(22 citation statements)
references
References 42 publications
0
22
0
Order By: Relevance
“…To enhance the network utility, Zhao et al optimized the charging vehicle route, the data rate, and the charging time together, and they presented multiple period iterative algorithms [16]. Considering that the energy consumption rate of nodes is dynamically changed, Yang et al proposed a real-time global charging scheme in wireless rechargeable sensor networks based on the actor-critic reinforcement learning algorithm [17]. Although existing studies have made many achievements from the perspectives of charging time, data transmission delay, network utility, etc., more work is still needed, as the mobility and communication performance of mobile charging vehicles are affected by terrain and other factors, while UAV assisted charging is rarely limited.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To enhance the network utility, Zhao et al optimized the charging vehicle route, the data rate, and the charging time together, and they presented multiple period iterative algorithms [16]. Considering that the energy consumption rate of nodes is dynamically changed, Yang et al proposed a real-time global charging scheme in wireless rechargeable sensor networks based on the actor-critic reinforcement learning algorithm [17]. Although existing studies have made many achievements from the perspectives of charging time, data transmission delay, network utility, etc., more work is still needed, as the mobility and communication performance of mobile charging vehicles are affected by terrain and other factors, while UAV assisted charging is rarely limited.…”
Section: Related Workmentioning
confidence: 99%
“…However, the drawbacks are also obvious. The path planning and communication quality of ground mobile chargers are susceptible to the terrain, and therefore it is difficult to replenish energy efficiently and to do so in a timely way [11][12][13][14][15][16][17].…”
Section: Introductionmentioning
confidence: 99%
“…However, these methods have very slow convergence [35]. To tackle this problem, the Deep Deterministic Policy Gradient abbreviated as DDPG is utilized to integrate both properties of policy-based and valuebased algorithms in order to deal with continuous and large state/action spaces [35], [36]. In this algorithm, there are two separated neural networks, an actor and a critic network with parameters ω and θ, respectively.…”
Section: ) Deep Deterministic Policy Gradient Deep Rl Algorithmmentioning
confidence: 99%
“…Cao et al [ 28 ] proposed a deep reinforcement learning-based on-demand charging algorithm to maximize the sum of rewards collected by the mobile charger in WRSN, which is subject to the energy capacity constraint on the mobile charger and the charging times of all sensor nodes. A novel charging scheme for dynamic WRSNs based on an actor–critic reinforcement learning algorithm was proposed by Yang et al [ 31 ], which aimed to maximize the charging efficiency while minimizing the number of dead sensors to prolong the network lifetime. The above works have made significant model innovation and algorithm innovation, yet they ignore the impact of sensor charging energy on the optimization performance.…”
Section: Introductionmentioning
confidence: 99%
“…The above works have made significant model innovation and algorithm innovation, yet they ignore the impact of sensor charging energy on the optimization performance. Although Yang et al [ 31 ] proposed a charging coefficient to constrain the upper charging energy threshold, they assumed that all sensors have a fixed charging coefficient during the scheduling, which cannot adjust according to the needs of the sensors. Specifically, the charging coefficient could directly determine the charging energy for the sensor.…”
Section: Introductionmentioning
confidence: 99%