2022
DOI: 10.1109/tkde.2020.3038109
|View full text |Cite
|
Sign up to set email alerts
|

CuWide: Towards Efficient Flow-Based Training for Sparse Wide Models on GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 42 publications
0
3
0
Order By: Relevance
“…Figure 3 illustrates the skew distribution of embedding update frequency on some popular workloads, including clickthrough rate prediction (i.e., Criteo), citation network (i.e., ogbnmag), and product co-purchasing network (i.e., Amazon). The top Existing research provides evidence that parameter updates from various embedding models exhibit a universal skewed distribution [35], such as recommendation models [8,24,52], LDA topic models [20,25,47,48] and graph learning models [33,40].…”
Section: Problems and Opportunitiesmentioning
confidence: 99%
See 1 more Smart Citation
“…Figure 3 illustrates the skew distribution of embedding update frequency on some popular workloads, including clickthrough rate prediction (i.e., Criteo), citation network (i.e., ogbnmag), and product co-purchasing network (i.e., Amazon). The top Existing research provides evidence that parameter updates from various embedding models exhibit a universal skewed distribution [35], such as recommendation models [8,24,52], LDA topic models [20,25,47,48] and graph learning models [33,40].…”
Section: Problems and Opportunitiesmentioning
confidence: 99%
“…Similar to TensorFlow [7], we use a static computation graph abstraction to organize all the operations in HET. All operators implemented by GPU kernels are scheduled into the GPU stream [35]. These operators will be launched and executed asynchronously to avoid blocking the CPU execution.…”
Section: Asynchronous Communication Invocationmentioning
confidence: 99%
“…Then a ranking model predicts a score for each candidate item and select the top items based on the estimated scores. This two-step procedure is widely adopted in the large scale industry recommender systems owing to its scalability and fast inference performance [1,2,3,4,5,6,7,8,9]. In this paper, we focus on the candidate generation stage [10], which is usually referred to as the top-N recommendation [11] in the academic area.…”
Section: Introductionmentioning
confidence: 99%