2020
DOI: 10.1109/tnse.2020.2966504
|View full text |Cite
|
Sign up to set email alerts
|

Data Collection Versus Data Estimation: A Fundamental Trade-Off in Dynamic Networks

Abstract: Index Terms-Networked systems, Partially observable Markov decision process, reinforcement learning, separation principle.1 The computational complexity of finding the optimal solution of the finitehorizon POMDP is PSpace-complete whereas that of MDP is P-complete [10]. 2 The notion of α-vectors was introduced in [13] to solve the finite-horizon POMDP, and was later enhanced by pruning dominated vectors [14], [15].

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3
3

Relationship

3
3

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 34 publications
0
5
0
Order By: Relevance
“…For more details on the connection with deep neural networks, the interested reader is referred to [32], [34].…”
Section: A Deep Chapman-kolmogorov (Dck) Equationmentioning
confidence: 99%
See 2 more Smart Citations
“…For more details on the connection with deep neural networks, the interested reader is referred to [32], [34].…”
Section: A Deep Chapman-kolmogorov (Dck) Equationmentioning
confidence: 99%
“…Note that the complexity of computing ˆ t (d t , γ t ) in time is exponential with respect to the number of agents. However, this computation can be carried out off-line by machine learning methods or circumvented by reinforcement learning techniques [34], [37]. In general, the exploration space of an arbitrarily asymmetric cost function grows exponentially with the number of agents while that of its deep state projection grows polynomially, according to Proposition 2, which is a considerable reduction in complexity.…”
Section: Infinite Horizon Discounted Costmentioning
confidence: 99%
See 1 more Smart Citation
“…Notice that z t is not necessarily the best possible estimate, but it is always measurable with respect to PDSS information structure. The interested reader is referred to [20] for more details on the above Kalman-like filter that emerges in the trade-off between data collection and data estimation.…”
Section: A Finite-model Strategy For Problemmentioning
confidence: 99%
“…Furthermore, we demonstrate in [9] that today's most common feed-forward deep neural networks (i.e., those with rectified linear unit activation function) may be viewed as a special case of deep structured teams, where layers are time steps and neurons are simple integrator agents whose goal is to collaborate in order to minimize a common loss (cost) function. For more applications of deep structured models, the reader is referred to reinforcement learning [8], [13], [14], nonzerosum game [12], [15], minmax optimization [17], leaderfollowers [18], [19], epidemics [20], smart grids [21], meanfield teams [22]- [25], and networked estimation [26], [27].…”
Section: Introductionmentioning
confidence: 99%