2022
DOI: 10.1007/978-3-031-19830-4_15
|View full text |Cite
|
Sign up to set email alerts
|

PACTran: PAC-Bayesian Metrics for Estimating the Transferability of Pretrained Models to Classification Tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 24 publications
0
6
0
Order By: Relevance
“…In the context of meta-learning, PAC-Bayesian theory is extensively studied to provide guarantees for generalization errors (Ding et al 2021;Farid and Majumdar 2021;.…”
Section: Hierarchical Pac-bayesian Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…In the context of meta-learning, PAC-Bayesian theory is extensively studied to provide guarantees for generalization errors (Ding et al 2021;Farid and Majumdar 2021;.…”
Section: Hierarchical Pac-bayesian Analysismentioning
confidence: 99%
“…The hierarchical analysis decomposes the entire problem into three tiers: task, subject, and curriculum, which allows us to construct the overall curriculum bound by combining the bounds from lower tiers. In addition, our theoretical contribution also makes two novel extensions to existing PAC-Bayes literature (Amit and Meir 2018; Rothfuss et al 2021;Ding et al 2021), including (i) deriving a bound on noisy meta-learning tasks and (ii) tackling the non i.i.d. task dependencies across different subjects.…”
Section: Introductionmentioning
confidence: 97%
“…To avoid exhaustive attempts on all pairs of source tasks and target tasks, TE provides efficient heuristics to exhibit the best-performing source task at a minor cost . Originated in the field of CV, a great number of TE approaches, including model-similarity-based methods (Dwivedi and Roig, 2019), label-comparison-based methods (Tran et al, 2019) and source features-based methods (Ding et al, 2022), etc., have been proposed in the past few years. To adapt such techniques to PLM selection for NLP tasks, Bassignana et al (2022) found the predictions of LogME can positively correlate with the true performances of candidate PLMs, and Vu et al (2022) exhibited the model similarity computed by soft prompts reflects the transfer performance across different models.…”
Section: Related Workmentioning
confidence: 99%
“…Model Similarity-based Methods DSE (Vu et al, 2020) ϕ(x), ψ(x) ✓ ✗ ✗ DDS (Dwivedi et al, 2020) ϕ(x), ψ(x) ✓ ✗ ✗ Training-free Methods MSC (Meiseles and Rokach, 2020) ϕ(x), y ✗ ✗ ✓ kNN (Puigcerver et al, 2021) ϕ(x), y ✗ ✗ ✓ PARC (Bolya et al, 2021) ϕ(x), y ✗ ✗ ✓ GBC ϕ(x), y ✗ ✗ ✓ Logistic (Kumari et al, 2022) ϕ(x), y ✗ ✓ ✓ H-score (Bao et al, 2019) ϕ(x), y ✗ ✗ ✓ Reg. H-score (Ibrahim et al, 2022) ϕ(x), y ✗ ✗ ✓ N LEEP ϕ(x), y ✗ ✗ ✓ TransRate (Huang et al, 2022) ϕ(x), y ✗ ✗ ✓ LogME (You et al, 2021) ϕ(x), y ✗ ✓ ✓ SFDA (Shao et al, 2022) ϕ(x), y ✗ ✓ ✓ PACTran (Ding et al, 2022) ϕ(x), y ✗ ✓ ✓ target classes by LR's test accuracy.…”
Section: Free Of Trainingmentioning
confidence: 99%
See 1 more Smart Citation