2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2020
DOI: 10.1109/ipdpsw50202.2020.00170
|View full text |Cite
|
Sign up to set email alerts
|

Data Parallel Large Sparse Deep Neural Network on GPU

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 18 publications
0
3
0
Order By: Relevance
“…Time series forecasting is the process of using a model to generate predictions (forecasts) for future events based on known past events [22,37]. There are several machine learning methods:s regression, classification, clustering [38], dimensionality reduction, ensemble methods, neural nets and deep learning [39], transfer learning, reinforcement learning, Natural Language Processing (NLP),word embeddings, etc. Regression is one of the predictive modeling techniques which analyzes the correlations between a target and independent variables.…”
Section: Time Series Forecastingmentioning
confidence: 99%
“…Time series forecasting is the process of using a model to generate predictions (forecasts) for future events based on known past events [22,37]. There are several machine learning methods:s regression, classification, clustering [38], dimensionality reduction, ensemble methods, neural nets and deep learning [39], transfer learning, reinforcement learning, Natural Language Processing (NLP),word embeddings, etc. Regression is one of the predictive modeling techniques which analyzes the correlations between a target and independent variables.…”
Section: Time Series Forecastingmentioning
confidence: 99%
“…The neural network parallelization process can occur during feed forward and backpropagation. It is because each node in a layer does not need information from other nodes in the same layer, so the process can run in parallel [34]. The first model only used the CPU in the modeling process because the processes were sequential.…”
Section: Build Prediction Modelmentioning
confidence: 99%
“…In this regard, one significant advantage of sparse models is that the sparse gradient communication is automatically at hand. Related work on parallelisation for sparse DNN is presented in (Sattar & Anfuzzaman (2020)) as a solution to the Sparse DNN Challenge posed by MIT/IEEE/Amazon. However, their work is focused on sparse neural networks created using RadiX-Net (Kepner & Robinett (2019)) which do not evolve the topology over time.…”
Section: Parallel Training Of Sparse Networkmentioning
confidence: 99%