ICC 2019 - 2019 IEEE International Conference on Communications (ICC) 2019
DOI: 10.1109/icc.2019.8761526
|View full text |Cite
|
Sign up to set email alerts
|

A Double Q-Learning Routing in Delay Tolerant Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 23 publications
(20 citation statements)
references
References 18 publications
0
18
0
Order By: Relevance
“…However, using Q-Learning in DTNs can also suffer a large penalty, because it can produce a positive bias by using the maximum value as the approximation of the maximum expected value. Therefore, we proposed DQLR [11] to solve the above problems. In the DQLR protocol, the Double Q -Learning algorithm is used to decouple the selection from the evaluation to obtain an unbiased estimation.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, using Q-Learning in DTNs can also suffer a large penalty, because it can produce a positive bias by using the maximum value as the approximation of the maximum expected value. Therefore, we proposed DQLR [11] to solve the above problems. In the DQLR protocol, the Double Q -Learning algorithm is used to decouple the selection from the evaluation to obtain an unbiased estimation.…”
Section: Related Workmentioning
confidence: 99%
“…But, the DTBR protocol takes the maximum action value as the optimal action, which may be obscured by overestimation. Therefore, in our previous work, the Double Q-Learning Routing (DQLR) protocol was proposed [11], which adopts the Double Q-Learning algorithm to obtain an unbiased estimation and improve the performance of message delivery. However, in real scenarios of DTNs, the characteristics of nodes (e.g., node activity, contact interval, and movement speed) are complex, dynamic, and uncertain, which will affect the performance of routing protocols.…”
Section: Introductionmentioning
confidence: 99%
“…The situation is quite similar to in SSA-MAC: the nano-nodes have to harvest energy and wait for their own slots, which could also cause latency and intermittent connectivity. For example, a double Q-learning routing algorithm is proposed for DTN to improve the connective efficiency, while balancing the routing performance and cost [45]. In [46], reliable energy-aware routing protocol for DTN is proposed to address the energy depletion of nodes.…”
Section: The Proposed Ssa-mac Protocolmentioning
confidence: 99%
“…Double Q-Learning Routing(DQLR) [22], selects the next hop in a distributed way. DQLR decouples the selection and evaluation with two value functions ie., the double Q-Learning functions.…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…Bayesian Classifier [8], [9], [10] K-Means Clustering [2] Principal Componenet Analysis [18] Q-Learning [13], [22], [25] [ 14] high trust value towards the destination, then that intermediate node is chosen for routing. If the intermediate nodes have the same trust value, then the routing decision is based on the latest connection time.…”
Section: A Routingmentioning
confidence: 99%