2019
DOI: 10.48550/arxiv.1906.07697
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fast and Flexible Multi-Task Classification Using Conditional Neural Adaptive Processes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 7 publications
(27 citation statements)
references
References 0 publications
0
27
0
Order By: Relevance
“…Meta-models. Meta-models include methods based on MAML [93], ProtoNets [94] and auxiliary nets predicting task-specific parameters [95][96][97][98]. These methods are tied to a particular architecture and need to be trained from scratch if it is changed.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Meta-models. Meta-models include methods based on MAML [93], ProtoNets [94] and auxiliary nets predicting task-specific parameters [95][96][97][98]. These methods are tied to a particular architecture and need to be trained from scratch if it is changed.…”
Section: Related Workmentioning
confidence: 99%
“…Denil et al [113] train a model that can predict a fraction of network parameters given other parameters requiring to retrain the model for each new architecture. Bertinetto et al [98] train a model that predicts parameters given a new few-shot task similarly to [18,96], and the model is also tied to a particular architecture. The HyperGAN [114] allows to generate an ensemble of trained parameters in a computationally efficient way, but as the aforementioned works is constrained to a particular architecture.…”
Section: Appendixmentioning
confidence: 99%
“…Another setting where FiLM layers have been shown effective is few-shot learning. This is the case, for instance, of TADAM (Oreshkin et al, 2018), CNAPs (Requeima et al, 2019), and CAVIA (Zintgraf et al, 2019) where FiLM layers are used to enable adapting a global cross-task model to particular tasks. Moreover, the few-shot classification setting under domain shift is tackled in (Tseng et al, 2020), where feature-wise transformations are used as a means to diversify data and artificially create new domains at training time.…”
Section: Conditional Modelingmentioning
confidence: 99%
“…where we have split Z n into context (Z n,c ) and target (Z n,t ) sets. This is standard practice in both the NP (Garnelo et al, 2018a;b) and meta-learning settings (Finn et al, 2017; and relates to neural auto-regressive models (Requeima et al, 2019). Practically, stochastic gradient descent methods (Bottou, 2010) can be used to perform the optimization.…”
Section: Convolutional Conditional Neural Processesmentioning
confidence: 99%
“…A key component of NPs is the embedding of context sets Z into a representation space through an encoder Z → E(Z), which is achieved using a deep set function approximator (Zaheer et al, 2017). This simple model specification allows NPs to be used for (i) meta-learning (Thrun & Pratt, 2012;Schmidhuber, 1987), since predictions can be generated on the fly from new context sets at test time; and (ii) multi-task or transfer learning (Requeima et al, 2019), since they provide a natural way of sharing information between data sets. Moreover, conditional NPs (CNPs; Garnelo et al, 2018a), a deterministic variant of NPs, can be trained in a particularly simple way with maximum likelihood learning of the parameters θ, which mimics how the system is used at test time, resulting in strong performance .…”
Section: Introductionmentioning
confidence: 99%