2014 IEEE Hot Chips 26 Symposium (HCS) 2014
DOI: 10.1109/hotchips.2014.7478821
|View full text |Cite
|
Sign up to set email alerts
|

SDA: Software-defined accelerator for large-scale DNN systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
46
0
1

Year Published

2015
2015
2022
2022

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 65 publications
(48 citation statements)
references
References 0 publications
1
46
0
1
Order By: Relevance
“…This is shown, e.g. by the decision by Baidu to accelerate its deep learning models for image search by using FPGAs [3]. Similarly, Microsoft, after years of research to accelerate its Bing search engine, is now also looking into how to accelerate deep learning models through FPGAs [4].…”
Section: B Overview Of Openclmentioning
confidence: 99%
See 1 more Smart Citation
“…This is shown, e.g. by the decision by Baidu to accelerate its deep learning models for image search by using FPGAs [3]. Similarly, Microsoft, after years of research to accelerate its Bing search engine, is now also looking into how to accelerate deep learning models through FPGAs [4].…”
Section: B Overview Of Openclmentioning
confidence: 99%
“…Modern FPGAs have the ability to provide sufficient processing speed while consuming a fraction of the power consumed by high-end GPUs [2]. This is why several big data companies such as Microsoft, Baidu are exploring FPGA devices as accelerators rather than GPUs [3,4].…”
Section: Introductionmentioning
confidence: 99%
“…They report almost doubled throughput in search ranking at a cost of only 10% increased power consumption and 30% increased total cost of ownership. Baidu presented accelerated neural networks on FPGAs in the datacenter offering an order of magnitude better performance at minimal additional power cost [13]. In [14], the authors present FPGAs in the cloud as standalone resources, treated separately from standard (server) resources.…”
Section: Existing Workmentioning
confidence: 99%
“…However, using such an index does not remove all the file system overhead from the IO-path. It is well known that permission checks of file systems increase latency in the IOpath by 10-15µs [13,43]. A combined memory and storagelevel indirection layer would not only eliminate a level of indirection but would also perform all the necessary checks efficiently by using the protection bits in the page tables.…”
Section: Challenges For a Combined Indirection Layermentioning
confidence: 99%