1990
DOI: 10.1209/0295-5075/11/6/001
|View full text |Cite
|
Sign up to set email alerts
|

A Convergence Theorem for Sequential Learning in Two-Layer Perceptrons

Abstract: We consider a Perceptron with N i input units, one output and a yet unspecified number of hidden units. This Perceptron must be able to learn a given but arbitrary set of input-output examples. By sequential learning we mean that groups of patterns, pertaining to the same class, are sequentially separated from the rest by successively adding hidden units until the remaining patterns are all in the same class. We prove that the internal representations obtained by such procedures are linearly separable. Prelimi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
64
0

Year Published

1995
1995
2014
2014

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 114 publications
(64 citation statements)
references
References 9 publications
0
64
0
Order By: Relevance
“…Sirat and Nadal proposed a similar algorithm in (Sirat & Nadal, 1990). Other interesting constructive techniques are sequential learning (Marchand, Golea, & Rújan, 1990), the patch algorithm (Barkema, Andree, & Taal, 1993), the oilspot algorithm (Frattale-Mascioli & Martinelli, 1995) and the techniques presented in (Muselli, 1995;Rújan & Marchand, 1989). Mezard and Nadal in (Mezard & Nadal, 1989), proposed a tiling algorithm which starts by training a single unit on the whole training set.…”
Section: Discussionmentioning
confidence: 99%
“…Sirat and Nadal proposed a similar algorithm in (Sirat & Nadal, 1990). Other interesting constructive techniques are sequential learning (Marchand, Golea, & Rújan, 1990), the patch algorithm (Barkema, Andree, & Taal, 1993), the oilspot algorithm (Frattale-Mascioli & Martinelli, 1995) and the techniques presented in (Muselli, 1995;Rújan & Marchand, 1989). Mezard and Nadal in (Mezard & Nadal, 1989), proposed a tiling algorithm which starts by training a single unit on the whole training set.…”
Section: Discussionmentioning
confidence: 99%
“…when it is less than about 75%, the PLR leads to the best or the second best generalization performance. However, there is no possibility of direct comparison of the proposed method to the previous works in terms of the generalization performance, since the most similar methods, those also employing the same data sets [4,5,22,23], do not declare their generalization performances. The generalization performance of the proposed cascade network is discussed in Section 4.5.…”
Section: Algorithmmentioning
confidence: 99%
“…It was proven in [4] that all samples can correctly be classified by the SLA in a 2-layer structure, where the outputs of the first layer become linearly separable and the second layer consists of just 1 neuron. However, the work in [4] dealt with Boolean inputs only. Another algorithm, based on SLA, is the constructive algorithm for real-valued examples (CARVE) [5], which extends the SLA from * Correspondence: ibrahim.genc@medeniyet.edu.tr Boolean inputs to real-valued input cases; it uses a convex hull method for the determination of the network weights instead of the PLR.…”
Section: Introductionmentioning
confidence: 99%
“…Sequential algorithm [7], instead of training neurons to classify a maximal subset of patterns, it trains neurons to sequentially exclude patterns belonging to one class from other. When all patterns are excluded, the internal representation of patterns in hidden layer is linearly separable.…”
Section: A Constructive Neural Network Algorithmsmentioning
confidence: 99%