2008
DOI: 10.1007/s10479-008-0424-0
|View full text |Cite
|
Sign up to set email alerts
|

Analysis of the consistency of a mixed integer programming-based multi-category constrained discriminant model

Abstract: Classification is concerned with the development of rules for the allocation of observations to groups, and is a fundamental problem in machine learning. Much of previous work on classification models investigates two-group discrimination. Multi-category classification is less-often considered due to the tendency of generalizations of two-group models to produce misclassification rates that are higher than desirable. Indeed, producing "good" two-group classification rules is a challenging task for some applica… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
19
0

Year Published

2008
2008
2018
2018

Publication Types

Select...
4
4
1

Relationship

2
7

Authors

Journals

citations
Cited by 38 publications
(20 citation statements)
references
References 23 publications
1
19
0
Order By: Relevance
“…Traditional SVM is tuned for the same parameter values except for the loss function. Classification trees are tuned for the split criterion (Gini or information) and k-nearest neighbor is tuned for k (1,3,4,7,9) and distance function (L 1 and L 2 ). Random forests and logistic regression are used with default settings for all tests.…”
Section: Comparisons With Other Classifiersmentioning
confidence: 99%
See 1 more Smart Citation
“…Traditional SVM is tuned for the same parameter values except for the loss function. Classification trees are tuned for the split criterion (Gini or information) and k-nearest neighbor is tuned for k (1,3,4,7,9) and distance function (L 1 and L 2 ). Random forests and logistic regression are used with default settings for all tests.…”
Section: Comparisons With Other Classifiersmentioning
confidence: 99%
“…Steinwart [36] proves that SVM with the traditional hinge loss is universally consistent. Brooks and Lee [7] prove that an integer-programming based method for constrained discrimination, a generalization of the classification problem, is consistent. This paper presents new integer programming formulations for SVM with the ramp loss and hard margin loss that accommodate the use of nonlinear kernel functions and the quadratic margin term.…”
Section: Introductionmentioning
confidence: 95%
“…The DAMIP classification model, a general-purpose, optimization-based, predictive modeling framework, has proven to be a very powerful supervised learning classification approach in predicting various biomedical and clinical phenomena [22][23][24] due to the universal consistency of the resulting classification rules and their ability to classify with high accuracy even among small training sets. 25 Fifty percent of the practices were randomly selected as the training set, 7 practices in group 1 and 9 in group 2. The DAMIP model was then applied to this training set to develop the prediction rule and to obtain an unbiased estimate of correct classification.…”
Section: Classification Analysesmentioning
confidence: 99%
“…6 and 7, surrogate functions provide a poor trade-off between accuracy and sparsity. Convex surrogate loss functions, for instance, produce models that do not attain the best learning-theoretic guarantee on predictive accuracy and are not robust to outliers (Li and Lin 2007;Brooks and Lee 2010;Nguyen and Sanner 2013). Similarly, 1 -regularization is only guaranteed to recover the correct sparse solution (i.e., the one that minimizes the 0 -norm) under restrictive conditions that are rarely satisfied in practice (Zhao and Bin 2007;Liu and Zhang 2009).…”
Section: Sparse Linear Classification Modelsmentioning
confidence: 99%