2022 58th Annual Allerton Conference on Communication, Control, and Computing (Allerton) 2022
DOI: 10.1109/allerton49937.2022.9929374
|View full text |Cite
|
Sign up to set email alerts
|

Understanding the Generalization Power of Overfitted NTK Models: 3-layer vs. 2-layer (Extended Abstract)

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 1 publication
0
1
0
Order By: Relevance
“…A major challenge here is that a single pre-trained model is unable to generalize well. We approached such a problem by leveraging a popular meta-learning framework of model-agnostic meta-learning (MAML), for which we first made several innovative contributions, including characterizing the optimization performance (Ji, Yang, and Liang 2021), understanding the statistical convergence guarantee (Collins et al 2022), and characterizing the generalization performance in the overparameterized regime (Ju, Liang, and Shroff 2023). Equipped with these powerful tools, we further demonstrated the advantages of MAML over traditional deep-learning techniques wherein a model retrained in the unseen test environment (i) uses a fraction of the data compared to classical retraining, which, in turn, simplifies data collection and storage, and (ii) results in equal or higher accuracy in optimal beam selection compared to the case when the new environment dataset is fully available during initial training.…”
Section: Meta-learning For Autonomous Transportationmentioning
confidence: 99%
“…A major challenge here is that a single pre-trained model is unable to generalize well. We approached such a problem by leveraging a popular meta-learning framework of model-agnostic meta-learning (MAML), for which we first made several innovative contributions, including characterizing the optimization performance (Ji, Yang, and Liang 2021), understanding the statistical convergence guarantee (Collins et al 2022), and characterizing the generalization performance in the overparameterized regime (Ju, Liang, and Shroff 2023). Equipped with these powerful tools, we further demonstrated the advantages of MAML over traditional deep-learning techniques wherein a model retrained in the unseen test environment (i) uses a fraction of the data compared to classical retraining, which, in turn, simplifies data collection and storage, and (ii) results in equal or higher accuracy in optimal beam selection compared to the case when the new environment dataset is fully available during initial training.…”
Section: Meta-learning For Autonomous Transportationmentioning
confidence: 99%