2015
DOI: 10.1007/978-3-319-27284-9_8
|View full text |Cite
|
Sign up to set email alerts
|

Why Some Families of Probability Distributions Are Practically Efficient: A Symmetry-Based Explanation

Abstract: Out of many possible families of probability distributions, some families turned out to be most efficient in practical situations. Why these particular families and not others? To explain this empirical success, we formulate the general problem of selecting a distribution with the largest possible utility under appropriate constraints. We then show that if we select the utility functional and the constraints which are invariant under natural symmetries -shift and scaling corresponding to changing the starting … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
3
0

Year Published

2016
2016
2017
2017

Publication Types

Select...
4

Relationship

3
1

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 29 publications
0
3
0
Order By: Relevance
“…It turns out that entropy is not the only functional with the above scale-invariance properties. All such scale-invariant functions have been described [5,6]. In addition to entropy, we can also have ∫ ln(ρ(x)) dx and ∫ (ρ(x)) q dx for some q.…”
Section: Alternative Symmetry-based Explanationmentioning
confidence: 99%
“…It turns out that entropy is not the only functional with the above scale-invariance properties. All such scale-invariant functions have been described [5,6]. In addition to entropy, we can also have ∫ ln(ρ(x)) dx and ∫ (ρ(x)) q dx for some q.…”
Section: Alternative Symmetry-based Explanationmentioning
confidence: 99%
“…It turns out that entropy is not the only functional with the above scale-invariance properties. All such scale-invariant functions have been described [5,6]. In addition to entropy, we can also have For scale-invariant generalizations of entropy, we get Student distribution.…”
Section: Alternative Symmetry-based Explanationmentioning
confidence: 99%
“…Traditionally in probability theory, when we only have partial knowledge about the probability distribution, we select a distribution for which the entropy − ∫ ρ(x) · ln(ρ(x)) dx attains the largest possible value (see, e.g., [3]), or, equivalently, for which the integral ∫ ρ(x) · ln(ρ(x)) dx attains the smallest possible value. It is worth mentioning that, in general, if we assume that the criterion for selecting a probability distribution is scale-invariant (in some reasonable sense), then this criterion is equivalent to optimizing either entropy, or generalized entropy ∫ ln(ρ(x)) dx or ∫ ρ α (x) dx, for some α > 0; see, e.g., [5]. Our analysis shows that the generalized entropy corresponding to α = 2 and α = 3 describes mean-squared robustness.…”
mentioning
confidence: 99%