2010
DOI: 10.1109/tnn.2010.2049580
|View full text |Cite
|
Sign up to set email alerts
|

On the Selection of Weight Decay Parameter for Faulty Networks

Abstract: Abstract-The weight-decay technique is an effective approach to handle overfitting and weight fault. For fault-free networks, without an appropriate value of decay parameter, the trained network is either overfitted or underfitted. However, many existing results on the selection of decay parameter focus on fault-free networks only. It is well known that the weight-decay method can also suppress the effect of weight fault. For the faulty case, using a test set to select the decay parameter is not practice becau… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
9
0
1

Year Published

2011
2011
2018
2018

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 36 publications
(11 citation statements)
references
References 40 publications
1
9
0
1
Order By: Relevance
“…Many regularizing methods focus on optimizing network target function during the training [1], [4]- [7]. One commonly used approach is introducing training criterion that discourages converging the network into building a complicated solution and suppresses the number of effective weights by directing the learning algorithm towards a solution with the least possible amount of zero weights such as weight decay [8]. Other approaches encourage less co-adapting between units and reduces overfitting such as Dropout [3].…”
Section: Introductionmentioning
confidence: 99%
“…Many regularizing methods focus on optimizing network target function during the training [1], [4]- [7]. One commonly used approach is introducing training criterion that discourages converging the network into building a complicated solution and suppresses the number of effective weights by directing the learning algorithm towards a solution with the least possible amount of zero weights such as weight decay [8]. Other approaches encourage less co-adapting between units and reduces overfitting such as Dropout [3].…”
Section: Introductionmentioning
confidence: 99%
“…La función general de regularización está dada por la minimización del riesgo total [15] (ecuación 1): La descomposición de pesos es una de las estrategias de regularización más utilizadas en la literatura [16], dado que su implementación es computacionalmente sencilla, no depende de parámetros adicionales y permite mejorar la capacidad de generalización de la red neuronal. Esta técnica fue propuesta por [8] y también trabajada por [17] bajo el nombre de regresión de borde (ridge regression).…”
Section: Descomposición De Pesos (Wd)unclassified
“…For small S b , 1 we can assume thatw is close to w. The Taylor series expansions of f (·, ·) and g(·, ·) are given by , w(t)) + G(x t , w(t))(b ⊗ w(t)). (11) Putting (10) and (11) into (9), and using (4), we can show that h(w(t)) in (8) is given by…”
Section: A Multiplicative Weight Noisementioning
confidence: 99%
“…It is normally required in these applications that delayed HNNs (DHNNs) be globally asymptotically stable. In the past decade a large number of publications on the issue of stability of DHNN have appeared (see [7]- [11] and the cited references therein).…”
Section: Introductionmentioning
confidence: 99%