2020 IEEE Conference on Control Technology and Applications (CCTA) 2020
DOI: 10.1109/ccta41146.2020.9206397
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning in Deep Structured Teams: Initial Results with Finite and Infinite Valued Features

Abstract: In this paper, we consider Markov chain and linear quadratic models for deep structured teams with discounted and time-average cost functions under two non-classical information structures, namely, deep state sharing and no sharing. In deep structured teams, agents are coupled in dynamics and cost functions through deep state, where deep state refers to a set of orthogonal linear regressions of the states. In this article, we consider a homogeneous linear regression for Markov chain models (i.e., empirical dis… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
6

Relationship

3
3

Authors

Journals

citations
Cited by 6 publications
(8 citation statements)
references
References 21 publications
0
8
0
Order By: Relevance
“…Following the terminology of deep structured models [8], [11], aggregate variables (linear regressions) defined in ( 1) and ( 3) are called deep variables due to the fact that their evolutions across time horizon are similar to feed-forward deep neural networks. Hence, for ease of reference, we refer to xt as deep state at time t in the sequel.…”
Section: Problem Formulationmentioning
confidence: 99%
See 4 more Smart Citations
“…Following the terminology of deep structured models [8], [11], aggregate variables (linear regressions) defined in ( 1) and ( 3) are called deep variables due to the fact that their evolutions across time horizon are similar to feed-forward deep neural networks. Hence, for ease of reference, we refer to xt as deep state at time t in the sequel.…”
Section: Problem Formulationmentioning
confidence: 99%
“…In the perfect sharing and deep state sharing, the certainty equivalence principle simplifies the analysis and results in two standard decoupled Riccati equations [8], [11]. In the imperfect observation case, however, the certainty equivalence principle does not hold.…”
Section: Lemma 6 (Relations Between Primitive Random Variables)mentioning
confidence: 99%
See 3 more Smart Citations