2022
DOI: 10.48550/arxiv.2203.12533
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pathways: Asynchronous Distributed Dataflow for ML

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 0 publications
0
5
0
1
Order By: Relevance
“…The development of distributed training techniques has substantially accelerated the pace of training larger models [14,29]. In Nebula-I, the training environment contains two parts, i.e.…”
Section: Parallization Layermentioning
confidence: 99%
“…The development of distributed training techniques has substantially accelerated the pace of training larger models [14,29]. In Nebula-I, the training environment contains two parts, i.e.…”
Section: Parallization Layermentioning
confidence: 99%
“…Переживание-воображение-наблюдения-объяснения и ментальная самоорганизация (кодопоезис) обуславливают «экстремальное обобщение», которое до настоящего времени отсутствует у машинного интеллекта (Imagination: power of abstract modeling of hypothetical situations; human cognition is capable of extreme generalization, quickly adapting to radically novel situations [109]; новые системы и исследовательские идеи машинного обучения обсуждаются в работе [74]).…”
Section: к вопросу о концепции «сильного интеллекта»unclassified
“…PaLM [10], aka. Pathways Language Model, is a densely-activated decoder-only transformer language model trained using Pathways [15], a large-scale ML accelerator orchestration system that enables highly efficient training across TPU pods. At the time of release, PaLM 540B achieved breakthrough performance on a suite of multi-step reasoning tasks [10].…”
Section: Modelsmentioning
confidence: 99%