2023
DOI: 10.32604/csse.2023.034710
|View full text |Cite
|
Sign up to set email alerts
|

A Novel Mixed Precision Distributed TPU GAN for Accelerated Learning Curve

Abstract: Deep neural networks are gaining importance and popularity in applications and services. Due to the enormous number of learnable parameters and datasets, the training of neural networks is computationally costly. Parallel and distributed computation-based strategies are used to accelerate this training process. Generative Adversarial Networks (GAN) are a recent technological achievement in deep learning. These generative models are computationally expensive because a GAN consists of two neural networks and tra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 17 publications
(18 reference statements)
0
2
0
Order By: Relevance
“…The impact of the multinode on spark and GPU was examined for medical use cases [35], [38]. The impact of multi-node TPU on GAN for double precision was developed but lacked deployment and did not address the current bottleneck issues in TPU [34]. The proposed model is superior to the existing works since it uses multi-GPU GAN implementation addressing the bottleneck issues, ensures the deployment and continuous retraining of the model, and makes it usable for real-time applications.…”
Section: Key Takeawaysmentioning
confidence: 99%
“…The impact of the multinode on spark and GPU was examined for medical use cases [35], [38]. The impact of multi-node TPU on GAN for double precision was developed but lacked deployment and did not address the current bottleneck issues in TPU [34]. The proposed model is superior to the existing works since it uses multi-GPU GAN implementation addressing the bottleneck issues, ensures the deployment and continuous retraining of the model, and makes it usable for real-time applications.…”
Section: Key Takeawaysmentioning
confidence: 99%
“…Utilizing high-performance hardware isn't the only option; one potential alternative is to parallelize and provide DNN training operations on many nodes instead. Under these circumstances, the amount of work that each node contributes to the calculation is minimal at best [ [3] , [4] , [5] , [6] , [7] ]. Despite this, the communication delay is a critical obstacle in distributed training due to the frequent communication requirements for delivering vast volumes of data across multiple computing nodes.…”
Section: Introductionmentioning
confidence: 99%