2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541)
DOI: 10.1109/ijcnn.2004.1380066
|View full text |Cite
|
Sign up to set email alerts
|

Softprop: softmax neural network backpropagation leaming

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…The class scores are computed in the fully connected layer. After that, the output of the softmax layer is an N -dimensional vector ( Rimer and Martinez, 2004 ), corresponding to the number of classes desired, and N is set to two classes (normal and pathological fetuses). In this work, the cross-entropy is adopted as the loss function in the softmax classification layer.…”
Section: Methodsmentioning
confidence: 99%
“…The class scores are computed in the fully connected layer. After that, the output of the softmax layer is an N -dimensional vector ( Rimer and Martinez, 2004 ), corresponding to the number of classes desired, and N is set to two classes (normal and pathological fetuses). In this work, the cross-entropy is adopted as the loss function in the softmax classification layer.…”
Section: Methodsmentioning
confidence: 99%
“…Dynamically updating the value of the error margin as training progresses is a straightforward extension to be evaluated. Softprop, a learning approach combining CB1 and SSE optimization during training by means of the error margin, has shown improvement over CB1 in a preliminary study (Rimer & Martinez, 2004) and a thorough analysis will be presented in future work. Using a value for the error margin local to each training instance and intelligently updating these values as training progresses also shows promise.…”
Section: Future Workmentioning
confidence: 97%
“…The value of µ can also be decreased, and remain negative as training is concluded to account for noisy outliers. A preliminary analysis of updating µ during training has shown promise (Rimer & Martinez, 2004).…”
Section: Increasing the Margin With Cb Trainingmentioning
confidence: 99%
“…Prior work has shown [8][9][10] that methods of calculating softer values for each training pattern based on the network's output vector improve generalization and reduce variance on classification problems over a corpus of benchmark learning problems. One of these, called lazy training or CB1, focuses on classification accuracy backpropagates an error signal through the network only when a pattern is misclassified.…”
Section: Motivation For Cb3mentioning
confidence: 99%
“…Classification-based (CB) error functions [9,10] are a relatively new method of training multi-layer perceptrons. The CB functions heuristically seek to directly minimize classification error by backpropagating network error only on misclassified patterns.…”
Section: Introductionmentioning
confidence: 99%