2013
DOI: 10.1016/j.inffus.2012.01.012
|View full text |Cite
|
Sign up to set email alerts
|

Objective priors from maximum entropy in data classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
21
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 39 publications
(21 citation statements)
references
References 28 publications
0
21
0
Order By: Relevance
“…Mérouane Debbah and Ralf Müller correctly describe these joint probabilities as a model with as many degrees of freedom as possible, which leaves free degrees for correlation to exist or not [4] (p.1674). This avoids the introduction of unjustified information [4] (p.1672) corresponding to the simple intuition behind PME: when updating your probabilities, waste no useful information and do not gain information unless the evidence compels you to gain it (see [4] (p.1685f), [5] (p.376), [6,7], [8] (p.186)). The principle comes with its own formal apparatus, not unlike probability theory itself: Shannon's information entropy [9], the Kullback-Leibler divergence (see [10,11], [12] (p.308ff), [13] (p.262ff)), the use of Lagrange multipliers (see [3] (p.409ff), [12] (p.327f), [13] (p.281)), and the log-inverse relationship between information and probability (see [14][15][16][17]).…”
Section: Jeffrey's Updating Principle and The Principle Of Maximum Enmentioning
confidence: 99%
“…Mérouane Debbah and Ralf Müller correctly describe these joint probabilities as a model with as many degrees of freedom as possible, which leaves free degrees for correlation to exist or not [4] (p.1674). This avoids the introduction of unjustified information [4] (p.1672) corresponding to the simple intuition behind PME: when updating your probabilities, waste no useful information and do not gain information unless the evidence compels you to gain it (see [4] (p.1685f), [5] (p.376), [6,7], [8] (p.186)). The principle comes with its own formal apparatus, not unlike probability theory itself: Shannon's information entropy [9], the Kullback-Leibler divergence (see [10,11], [12] (p.308ff), [13] (p.262ff)), the use of Lagrange multipliers (see [3] (p.409ff), [12] (p.327f), [13] (p.281)), and the log-inverse relationship between information and probability (see [14][15][16][17]).…”
Section: Jeffrey's Updating Principle and The Principle Of Maximum Enmentioning
confidence: 99%
“…Process classification based on small values of n may be important in timecritical applications where an inference, or a decision, has to be made as soon as possible. In application where we have knowledge of the likelihoods, but we have no information about priors the classical use of uniform priors may lead to contradictory results [17]. Entropic priors, obtained from maximization of the joint entropy H(X 1:n , S), are to be considered "objective" and are…”
Section: Sequence Classificationmentioning
confidence: 99%
“…Entropic priors are derived from the maximization of model entropy and seem to be an excellent candidate for solving some of the inconsistencies related to model uncertainties [20,19,13,3,4,7,14]. We have presented a discussion on entropic priors in [17] and applied it to target classification [15], to graphical models [5] and to AR process classification [16]. Application of the Maximum Entropy principle [9] to prior determination qualifies entropic priors as a promising alternative to blind use of uniform priors.…”
Section: Introductionmentioning
confidence: 99%
“…As a result, the Bayesian and Neyman-Pearson approaches are ineffective due to the absence of complete prior information, and the Minimax approach achieves a poor performance since the available partial prior information is ignored. In order to utilize the partial information and to achieve a better performance, several studies have been conducted [8,[17][18][19][20][21][22][23][24][25], for example, maximum entropy (ME), Γ-minimax, restricted Bayes, and restricted Neyman-Pearson approaches, to name but a few. For instance, the ME method is utilized in [23] to translate the information contained in the known form of the likelihood into a prior distribution for Bayesian inference.…”
Section: Introductionmentioning
confidence: 99%