2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.00130
|View full text |Cite
|
Sign up to set email alerts
|

Can We Characterize Tasks Without Labels or Features?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 24 publications
0
3
0
Order By: Relevance
“…Neural Tangent Kernels. NTK were first introduced in [18] and developed in subsequent works [31][32][33][34]. NTK employ the gradient of kernels to describe the convergence behavior and to mimic the performance of (over-parameterized) DNN within a limit of infinite width.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Neural Tangent Kernels. NTK were first introduced in [18] and developed in subsequent works [31][32][33][34]. NTK employ the gradient of kernels to describe the convergence behavior and to mimic the performance of (over-parameterized) DNN within a limit of infinite width.…”
Section: Related Workmentioning
confidence: 99%
“…NTK employ the gradient of kernels to describe the convergence behavior and to mimic the performance of (over-parameterized) DNN within a limit of infinite width. TTK [34] uses a kernelized distance across the gradients of multiple random initialized networks to estimate the similarity over different tasks. While the ultra-wide networks do not apply to incremental learning, capturing the gradient flow in each incremental learning step could assist in coordinating the corresponding feature representations.…”
Section: Related Workmentioning
confidence: 99%
“…The similarity between tasks is the key point of selecting tasks. Researchers have proposed some methods to measure the distance between tasks including measuring method with probe network [8], [9] and measuring method with distribution [10]. The probe shows the optimal transport distance (or similarity) between 30000 training tasks with another 1 target task.…”
Section: Introductionmentioning
confidence: 99%