2022
DOI: 10.48550/arxiv.2212.09535
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

BLOOM+1: Adding Language Support to BLOOM for Zero-Shot Prompting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…LLMs possess multilingual capabilities that enable them to address language barriers, accommodate low-resource languages, and exhibit promising performance even on unseen languages (Yong et al, 2022). To enhance accessibility, the development and adoption of open-source multilingual models, such as BLOOM (Scao et al, 2022), should be encouraged, thereby facilitating the utilization of LLMs in educational applications across diverse linguistic contexts.…”
Section: Discussionmentioning
confidence: 99%
“…LLMs possess multilingual capabilities that enable them to address language barriers, accommodate low-resource languages, and exhibit promising performance even on unseen languages (Yong et al, 2022). To enhance accessibility, the development and adoption of open-source multilingual models, such as BLOOM (Scao et al, 2022), should be encouraged, thereby facilitating the utilization of LLMs in educational applications across diverse linguistic contexts.…”
Section: Discussionmentioning
confidence: 99%
“…While the BLOOM models were trained on data from 46 different languages, the training did not include Finnish. Prior work has investigated extending smaller BLOOM models to new languages not included during pretraining (Yong et al, 2022) and found parameter-efficient finetuning methods and (to a lesser degree) continued pretraining to be effective approaches. Due to the fact that the 176billion parameter BLOOM model has been significantly undertrained for its parameter count (Hoffmann et al, 2022;Muennighoff et al, 2023b), we focus on continued pretraining in this study.…”
Section: Modelsmentioning
confidence: 99%
“…Yle Archives of the national public broadcasting et al, 2022a). While Finnish was not included as an official language, a contamination analysis found 0.03% of ROOTS to be Finnish (Muennighoff et al, 2022). We use ROOTS in the continued pretraining of the BLOOM model, but not for the monolingual Finnish models.…”
Section: Data Sourcesmentioning
confidence: 99%
“…Large autoregressive language models (LLMs) such as GPT [3], ChatGPT [12], PaLM [4], or BLOOM [17] have the potential to address both of these shortcomings. Due to being pre-trained on huge amounts of text as well as due to emergent effects resulting from the model size [16], LLMs often have a better zero-shot performance compared to PLMs such as BERT and are also more robust concerning unseen examples [3].…”
Section: Introductionmentioning
confidence: 99%