2022
DOI: 10.48550/arxiv.2206.12131
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MVP: Multi-task Supervised Pre-training for Natural Language Generation

Abstract: Pre-trained language models (PLMs) have achieved notable success in natural language generation (NLG) tasks. Up to now, most of the PLMs are pre-trained in an unsupervised manner using large-scale general corpus. In the meanwhile, an increasing number of models pre-trained with less labeled data showcase superior performance compared to unsupervised models. Motivated by the success of supervised pre-training, we propose Multi-task superVised Pre-training (MVP) for natural language generation. For pre-training … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 69 publications
0
1
0
Order By: Relevance
“…Along with that, demands for a unified input data format have recently been raised with multi-task training for large language models (LLMs) (Sanh et al, 2021;Scao et al, 2022;Ouyang et al, 2022, inter alia). Some works have used simple data linearization techniques for converting structured data to a textual format, in order to align it with the format used for other tasks (Xie et al, 2022;Tang et al, 2022). However, linearizations are using custom preprocessing code, leading to discrepancies between individual works.…”
Section: Web Interfacementioning
confidence: 99%
“…Along with that, demands for a unified input data format have recently been raised with multi-task training for large language models (LLMs) (Sanh et al, 2021;Scao et al, 2022;Ouyang et al, 2022, inter alia). Some works have used simple data linearization techniques for converting structured data to a textual format, in order to align it with the format used for other tasks (Xie et al, 2022;Tang et al, 2022). However, linearizations are using custom preprocessing code, leading to discrepancies between individual works.…”
Section: Web Interfacementioning
confidence: 99%
“…Along with that, demands for a unified input data format have recently been raised with multi-task training for large language models (LLMs) (Sanh et al, 2022;Scao et al, 2022;Ouyang et al, 2022, inter alia). Some works have used simple data linearization techniques for converting structured data to a textual format, in order to align it with the format used for other tasks (Xie et al, 2022;Tang et al, 2022). However, linearizations are using custom preprocessing code, leading to discrepancies between individual works.…”
Section: Web Interfacementioning
confidence: 99%