2013
DOI: 10.1109/tvt.2013.2273945
|View full text |Cite
|
Sign up to set email alerts
|

Flexible, Portable, and Practicable Solution for Routing in VANETs: A Fuzzy Constraint Q-Learning Approach

Abstract: Vehicular ad hoc networks (VANETs) have been attracting interest for their potential uses in driving assistance, traffic monitoring, and entertainment systems. However, due to vehicle movement, limited wireless resources, and the lossy characteristics of a wireless channel, providing a reliable multihop communication in VANETs is particularly challenging. In this paper, we propose PFQ-AODV, which is a portable VANET routing protocol that learns the optimal route by employing a fuzzy constraint Q-learning algor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
77
0
1

Year Published

2015
2015
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 132 publications
(86 citation statements)
references
References 30 publications
(52 reference statements)
0
77
0
1
Order By: Relevance
“…(4), (5), (6), (7), (8), (10), (11),and (12) Calculate Steady state probablities by eq (13) Calculate freezing probability by eq. (14) Calculate updated transmission probability by eq.…”
Section: ) Channel Access Delaymentioning
confidence: 99%
See 1 more Smart Citation
“…(4), (5), (6), (7), (8), (10), (11),and (12) Calculate Steady state probablities by eq (13) Calculate freezing probability by eq. (14) Calculate updated transmission probability by eq.…”
Section: ) Channel Access Delaymentioning
confidence: 99%
“…The static quantification of channel quality is unable to incorporate the impact of vehicular environments considering the dynamic characteristics of the parameters directly affecting channel quality. Moreover, fuzzy-based techniques have been effectively applied for the quantification of parameters in dynamic environments in various domains [13].…”
mentioning
confidence: 99%
“…Q-learning is a recent form reinforcement learning that does not need a model of its environment and works by estimating the values of state-action pairs. It learns behavior through trial-and-error interactions with a dynamic environment, and has been employed for path selection [24]. The algorithm maintains a Q-value Q (s, a) in a table for every state-action pair.…”
Section: Improved Q-learning Mathematic Modelmentioning
confidence: 99%
“…The success of Q-Learning has lead to many applications, such as path planning [25,31], energy management [30], routing in vehicular ad-hoc networks [42], management of water resources [27], and production planning [10].…”
Section: Q-learningmentioning
confidence: 99%