2021
DOI: 10.48550/arxiv.2109.09701
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…Furthermore, the phobert-base model is the small architecture that is adapted to such a small dataset as the VieCap4H dataset, leading to a quick training time, which helps us conduct more experiments. We also try PhoBERT-large, BARTPho-syllable and BARTPho-word [19] pre-trained models, but it does not seem to operate well. The reason may be that the large architectures are not suitable for the small dataset as VieCap4H (contains 8032 samples).…”
Section: Language Embeddingmentioning
confidence: 99%
“…Furthermore, the phobert-base model is the small architecture that is adapted to such a small dataset as the VieCap4H dataset, leading to a quick training time, which helps us conduct more experiments. We also try PhoBERT-large, BARTPho-syllable and BARTPho-word [19] pre-trained models, but it does not seem to operate well. The reason may be that the large architectures are not suitable for the small dataset as VieCap4H (contains 8032 samples).…”
Section: Language Embeddingmentioning
confidence: 99%