2019 IEEE Conference on Control Technology and Applications (CCTA) 2019
DOI: 10.1109/ccta.2019.8920662
|View full text |Cite
|
Sign up to set email alerts
|

Feedback Control for Online Training of Neural Networks

Abstract: Convolutional neural networks (CNNs) are commonly used for image classification tasks, raising the challenge of their application on data flows. During their training, adaptation is often performed by tuning the learning rate. Usual learning rate strategies are time-based i.e. monotonously decreasing. In this paper, we advocate switching to a performance-based adaptation, in order to improve the learning efficiency. We present E (Exponential)/PD (Proportional Derivative)-Control, a conditional learning rate st… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
2
1

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(9 citation statements)
references
References 11 publications
0
9
0
Order By: Relevance
“…As Co-Teaching trains two models, we use two 56-layers ResNet. To speed up model convergence for RAD Slim, RAD Slim Limited, and Forward, we implement the E (Exponential)/PD (Proportional-Derivative)-Control [55] and Event-Based Control Learning rate [56] as learning rate schedule based on stochastic gradient descent (SGD) optimizer. Co-Teaching has its own learning rate scheduler.…”
Section: B Experimental Setupmentioning
confidence: 99%
“…As Co-Teaching trains two models, we use two 56-layers ResNet. To speed up model convergence for RAD Slim, RAD Slim Limited, and Forward, we implement the E (Exponential)/PD (Proportional-Derivative)-Control [55] and Event-Based Control Learning rate [56] as learning rate schedule based on stochastic gradient descent (SGD) optimizer. Co-Teaching has its own learning rate scheduler.…”
Section: B Experimental Setupmentioning
confidence: 99%
“…For the sake of simplicity the loss values are normalized with respect to the initial epoch loss value L(0). K P and K D are the proportional and derivative gain detailed in [11].…”
Section: A Event-based Learning Ratementioning
confidence: 99%
“…Note that the stability of CNN is ensured by E/PD, whose stability analysis is provided in [11]. Proposed event-based control does not introduce any instability because if e 1 = 0, which means the loss is decreasing, model is converging, and if e 1 = 1, the learning rate strategy returns to E/PD.…”
Section: A Event-based Learning Ratementioning
confidence: 99%
See 2 more Smart Citations