Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-1091
|View full text |Cite
|
Sign up to set email alerts
|

FastICARL: Fast Incremental Classifier and Representation Learning with Efficient Budget Allocation in Audio Sensing Applications

Abstract: Various incremental learning (IL) approaches have been proposed to help deep learning models learn new tasks/classes continuously without forgetting what was learned previously (i.e., avoid catastrophic forgetting). With the growing number of deployed audio sensing applications that need to dynamically incorporate new tasks and changing input distribution from users, the ability of IL on-device becomes essential for both efficiency and user privacy.However, prior works suffer from high computational costs and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
8
1

Relationship

3
6

Authors

Journals

citations
Cited by 11 publications
(4 citation statements)
references
References 21 publications
(29 reference statements)
0
4
0
Order By: Relevance
“…For other datasets (CIFAR-10, SVHN, GTSRB, GSC), we use variants of Mi-croNet architecture to construct pretrained models. To identify a high-performing and yet lightweight model to operate on embedded and mobile devices, we conduct a hyper-parameter search based on different variants of MicroNet (e.g., small, medium, large models), lightweight convolutional neural network (CNN) architectures [38], the number of convolutional filters. A basic convolutional layer consists of 3 × 3 convolution, batch normalization, and Rectified Linear Unit (ReLU).…”
Section: Performancementioning
confidence: 99%
“…For other datasets (CIFAR-10, SVHN, GTSRB, GSC), we use variants of Mi-croNet architecture to construct pretrained models. To identify a high-performing and yet lightweight model to operate on embedded and mobile devices, we conduct a hyper-parameter search based on different variants of MicroNet (e.g., small, medium, large models), lightweight convolutional neural network (CNN) architectures [38], the number of convolutional filters. A basic convolutional layer consists of 3 × 3 convolution, batch normalization, and Rectified Linear Unit (ReLU).…”
Section: Performancementioning
confidence: 99%
“…The first group of approaches is a regularization-based method [22,43,45,55] where regularization terms are added to the loss function to minimize changes to important weights of a model for previous tasks to prevent forgetting. Another group of approaches is a replay-based method [28] where model parameters are updated for learning a representation by using training data of the currently available classes, which is different from replay with exemplars-based method [23,29,40] where updating the model requires training data from the new class and also few training samples from earlier classes.…”
Section: Continual Learningmentioning
confidence: 99%
“…This leads to higher computational costs as the model expands and prohibits the utilization of compile-time optimizations on a fixed computation graph of the model. The last group of approaches among conventional CL includes rehearsal-based methods [7,8,26,49,63,67,74,81,98]. These prevent forgetting by replaying the saved rehearsal samples from earlier classes, typically leading to superior CL performance over the other methods at the cost of increased memory footprint.…”
Section: Introductionmentioning
confidence: 99%