2011 Wireless Advanced 2011
DOI: 10.1109/wiad.2011.5983301
|View full text |Cite
|
Sign up to set email alerts
|

Improved decentralized Q-learning algorithm for interference reduction in LTE-femtocells

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2012
2012
2023
2023

Publication Types

Select...
3
3
3

Relationship

2
7

Authors

Journals

citations
Cited by 52 publications
(31 citation statements)
references
References 3 publications
0
31
0
Order By: Relevance
“…where α = 0.5 is the player's willingness to learn from its environment, λ = 0.9 is the discount factor, and s ′ is the next state [34], [35]. Hereby, the agent's previous knowledge about the state-action pair (s, a) is represented by the first term in (9).…”
Section: B Q-learning Based Time-domain Icicmentioning
confidence: 99%
“…where α = 0.5 is the player's willingness to learn from its environment, λ = 0.9 is the discount factor, and s ′ is the next state [34], [35]. Hereby, the agent's previous knowledge about the state-action pair (s, a) is represented by the first term in (9).…”
Section: B Q-learning Based Time-domain Icicmentioning
confidence: 99%
“…Moreover, an improved effective initialization procedure was provided to overcome slow convergence the drawback of Q-learning. The RL approach based on Q-Learning framework has been also proposed to handle channel sharing between small cells deployed in a macro cell [41]. Automatic Neighbour Relation Management concerns the updates of neighbour cell relationships to facilitate easy handovers between base stations.…”
Section: Automatic Configuration Of Initial Radio Transmission Paramementioning
confidence: 99%
“…To overcome this drawback, we proposed an initialization procedure, which shows major convergence/performance enhancement [6]. Visiting one state for the first time not only the Q-value of a single state-action pair is updated, but estimates for the cost function for all other possible actions of the current state are added, so that the Qtable is rapidly initialized.…”
Section: Decentralized Q-learning With Improved Initializationmentioning
confidence: 99%