Twenty-First International Conference on Machine Learning - ICML '04 2004
DOI: 10.1145/1015330.1015339
|View full text |Cite
|
Sign up to set email alerts
|

Learning Bayesian network classifiers by maximizing conditional likelihood

Abstract: Bayesian networks are a powerful probabilistic representation, and their use for classification has received considerable attention. However, they tend to perform poorly when learned in the standard way. This is attributable to a mismatch between the objective function used (likelihood or a function thereof) and the goal of classification (maximizing accuracy or conditional likelihood). Unfortunately, the computational cost of optimizing structure and parameters for conditional likelihood is prohibitive. In th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
160
0
7

Year Published

2005
2005
2019
2019

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 212 publications
(171 citation statements)
references
References 22 publications
4
160
0
7
Order By: Relevance
“…Instead of optimizing the loglikelihood, discriminative approaches try to maximize the conditional likelihood and may produce better class probability estimates [14], [24], [3]. A discriminant version of SNODE can be developed by maximizing the conditional likelihood; however, the corresponding optimization problem may be more difficult than the current one, since the conditional likelihood does not decompose into separate terms for each function to be learned.…”
Section: Discussionmentioning
confidence: 99%
“…Instead of optimizing the loglikelihood, discriminative approaches try to maximize the conditional likelihood and may produce better class probability estimates [14], [24], [3]. A discriminant version of SNODE can be developed by maximizing the conditional likelihood; however, the corresponding optimization problem may be more difficult than the current one, since the conditional likelihood does not decompose into separate terms for each function to be learned.…”
Section: Discussionmentioning
confidence: 99%
“…A variety of multinomial classifiers have been proposed for handling an arbitrary number of independent attributes. Let us mention especially [45], [46], [34], semi-naive Bayesian classifiers [44], [19], tree-augmented naive Bayesian classifiers [32], k-dependence Bayesian classifiers [53], and Bayesian Network-augmented naive Bayesian classifiers [13].…”
Section: Appendix: Naive Bayesian Classifiersmentioning
confidence: 99%
“…There are large number of classifiers that are used to classify the data such as bayes, function, rule based and Tree etc. The goal of classification is to correctly predict the value of a designated discrete class variable, given a vector of predictors or attributes [5].…”
Section: Classificationmentioning
confidence: 99%