2020
DOI: 10.48550/arxiv.2011.04144
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu

Abstract: We provide finite sample guarantees for the classical Chow-Liu algorithm (IEEE Trans. Inform. Theory, 1968) to learn a tree-structured graphical model of a distribution. For a distribution P on Σ n and a tree T on n nodes, we say T is an ε-approximate tree for P if there is a T -structured distribution Q such that D(P Q) is at most ε more than the best possible tree-structured distribution for P . We show that if P itself is tree-structured, then the Chow-Liu algorithm with the plug-in estimator for mutual inf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(6 citation statements)
references
References 41 publications
(17 reference statements)
0
6
0
Order By: Relevance
“…Challenge 1: Failure of Chow-Liu. The main difficulty in proving our result is that the classical Chow-Liu algorithm, which is well-known to be optimal for structure learning and was recently shown to be optimal for total variation [6,13], does not work for learning to within locTV.…”
Section: Techniques and Main Ideasmentioning
confidence: 95%
See 4 more Smart Citations
“…Challenge 1: Failure of Chow-Liu. The main difficulty in proving our result is that the classical Chow-Liu algorithm, which is well-known to be optimal for structure learning and was recently shown to be optimal for total variation [6,13], does not work for learning to within locTV.…”
Section: Techniques and Main Ideasmentioning
confidence: 95%
“…To prove this theorem we construct a distribution that is close to a tree Ising model in local total variation, such that even when given population (infinite sample) values for the correlations the Chow-Liu algorithm makes local errors that accumulate at global scales. This issue does not arise in the analysis of [6,13], since they assume a model is -close in total variation to a tree model, and thus the input distribution is extremely close to a tree model in a global sense. Furthermore, in converting this to a finite sample result they assume access to at least n log n samples, which results in extremely small local estimation errors that do not appreciably accumulate.…”
Section: Techniques and Main Ideasmentioning
confidence: 99%
See 3 more Smart Citations