2023
DOI: 10.1007/978-3-031-23190-2_3
|View full text |Cite
|
Sign up to set email alerts
|

Improving Pre-trained Language Models

Abstract: This chapter describes a number of different approaches to improve the performance of Pre-trained Language Models (PLMs), i.e. variants of BERT, autoregressive language models similar to GPT, and sequence-to-sequence models like Transformers. First we may modify the pre-training tasks to learn as much as possible about the syntax and semantics of language. Then we can extend the length of the input sequence to be able to process longer inputs. Multilingual models are simultaneously trained with text in differe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 95 publications
0
0
0
Order By: Relevance
“…Sentiment analysis on film reviews using transformers (Putri et al, 2022;Bacco et al, 2021). Transformer introduces BERT and GPT with encoder and decoder (Ghojogh, 2020;Paaß & Giesselbach, 2023).…”
Section: Introductionmentioning
confidence: 99%
“…Sentiment analysis on film reviews using transformers (Putri et al, 2022;Bacco et al, 2021). Transformer introduces BERT and GPT with encoder and decoder (Ghojogh, 2020;Paaß & Giesselbach, 2023).…”
Section: Introductionmentioning
confidence: 99%