Proceedings of the 23rd International Conference on Machine Learning - ICML '06 2006
DOI: 10.1145/1143844.1143970
|View full text |Cite
|
Sign up to set email alerts
|

Totally corrective boosting algorithms that maximize the margin

Abstract: We consider boosting algorithms that maintain a distribution over a set of examples. At each iteration a weak hypothesis is received and the distribution is updated. We motivate these updates as minimizing the relative entropy subject to linear constraints. For example AdaBoost constrains the edge of the last hypothesis w.r.t. the updated distribution to be at most γ = 0. In some sense, AdaBoost is "corrective" w.r.t. the last hypothesis. A cleaner boosting method is to be "totally corrective": the edges of al… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

1
67
0

Year Published

2006
2006
2020
2020

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 83 publications
(68 citation statements)
references
References 20 publications
1
67
0
Order By: Relevance
“…It uses coordinate descent, however, and is therefore not fully (totally) corrective. Totally corrective boosting algorithms, like LPBoost [3], TotalBoost [18] and those proposed in [13], update the coefficients of all previously selected weak learners at each iteration. The fully corrective boosting algorithms thus require significantly fewer training iterations to achieve convergence [13] and result in smaller, and therefore more efficient, ensemble classifiers.…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…It uses coordinate descent, however, and is therefore not fully (totally) corrective. Totally corrective boosting algorithms, like LPBoost [3], TotalBoost [18] and those proposed in [13], update the coefficients of all previously selected weak learners at each iteration. The fully corrective boosting algorithms thus require significantly fewer training iterations to achieve convergence [13] and result in smaller, and therefore more efficient, ensemble classifiers.…”
Section: Introductionmentioning
confidence: 99%
“…In the proposed CGBoost, generally we do not require sophisticated convex solvers and only gradient descent methods like L-BFGS-B [19] are needed. Previous totally-corrective boosting algorithms [3,13,18] all solve the dual problems using convex optimization solvers. Besides the primal problems' much simpler structures, in most cases, the dual problems have many more variables than their corresponding primal problems.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations