2020
DOI: 10.48550/arxiv.2006.16823
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Technical Report: Auxiliary Tuning and its Application to Conditional Text Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(6 citation statements)
references
References 9 publications
0
5
0
Order By: Relevance
“…Side tuning [32] adds a side model that learns a residual on top of the original model. Similarly, [31] supplements the pre-trained TLM with an external model that shifts the output distribution. An important difference with our work is that we consider that the attribute model already exists in the TLM rather than using external models as in…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Side tuning [32] adds a side model that learns a residual on top of the original model. Similarly, [31] supplements the pre-trained TLM with an external model that shifts the output distribution. An important difference with our work is that we consider that the attribute model already exists in the TLM rather than using external models as in…”
Section: Related Workmentioning
confidence: 99%
“…Large and powerful language models [5] based on the Transformer architecture [28] (TLMs) achieve impressive performance [23,6]. However, such powerful models present several disadvantages: (i) they are difficult to train due to both the size of models and datasets, and the compute resources required 1 ; (ii) TLMs inherit and perpetuate biases that can have a negative social impact [1] and (iii) conditioning these models on concepts requires re-training [14] or using additional parameters [9,32,31], and being limited to very specific concepts.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations