2010 Ninth International Conference on Machine Learning and Applications 2010
DOI: 10.1109/icmla.2010.20
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Class Classification Using a New Sigmoid Loss Function for Minimum Classification Error (MCE)

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 5 publications
0
4
0
Order By: Relevance
“…Many studies tried to solve SVM with 0-1 loss function by replacing it with other smooth functions such as sigmoid function [22,23] , logistic function [24] , polynomial function [23] , and hyperbolic tangent function [25] . These functions, while yielding accurate result, still suffer from being computationally-expensive when solved as an optimization problem due to its non-convex nature.…”
Section: Multiclass Classification With Ramp Lossmentioning
confidence: 99%
“…Many studies tried to solve SVM with 0-1 loss function by replacing it with other smooth functions such as sigmoid function [22,23] , logistic function [24] , polynomial function [23] , and hyperbolic tangent function [25] . These functions, while yielding accurate result, still suffer from being computationally-expensive when solved as an optimization problem due to its non-convex nature.…”
Section: Multiclass Classification With Ramp Lossmentioning
confidence: 99%
“…In our work, the higher-order potentials are directly modeled by the cluster-based features with a sigmoid function. The sigmoid function is usually used as the activation function in many classification methods [44][45][46], which can be seen in Figure 6.…”
Section: Higher-order Potentialsmentioning
confidence: 99%
“…In our work, the higher-order potentials are directly modeled by the cluster-based features with a sigmoid function. The sigmoid function is usually used as the activation function in many classification methods [44][45][46], which can be seen in Figure 6. Before computing the higher-order energy of CRF defined in (9), the cluster-based features are normalized in [0,1] to balance the perception between features and classes.…”
Section: Higher-order Potentialsmentioning
confidence: 99%
“…This discriminant function is a smoothed approximation of the score difference between reference and decoding hypotheses and is usually used as a criterion for an objective function [17,42]. The relation between the true classification risk and the smoothed MCE loss function is recently studied in [43,44]. These studies revealed that the direct minimization of the classification error can be achieved if we could minimize the MCE loss function.…”
Section: I M C E -B a S E D D I S C R I M I N A T I V E T R A I Nmentioning
confidence: 99%