2022
DOI: 10.1109/mm.2021.3097700
|View full text |Cite
|
Sign up to set email alerts
|

Near-Memory Processing in Action: Accelerating Personalized Recommendation With AxDIMM

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 55 publications
(14 citation statements)
references
References 16 publications
0
9
0
Order By: Relevance
“…Many works from academia [2, 10-12, 15-23, 25, 31, 35-39, 48, 81-83, 85, 86, 90, 99, 104-112] and industry [34,[41][42][43][50][51][52][53][54] have shown the benefits of PnM and PuM for a wide range of workloads from different domains. However, fully adopting PIM in commercial systems is still very challenging due to the lack of tools and system support for PIM architectures across the computer architecture stack [4], which includes: (i) workload characterization methodologies and benchmark suites targeting PIM architectures; (ii) frameworks that can facilitate the implementation of complex operations and algorithms using the underlying PIM primitives (e.g., simple PIM arithmetic operations [19], bulk bitwise Boolean in-DRAM operations [83,84,92]); (iii) compiler support and compiler optimizations targeting PIM architectures; (iv) operating system support for PIM-aware virtual memory, memory management, data allocation and mapping; and (v) efficient data coherence and consistency mechanisms.…”
Section: Motivation and Problemmentioning
confidence: 99%
“…Many works from academia [2, 10-12, 15-23, 25, 31, 35-39, 48, 81-83, 85, 86, 90, 99, 104-112] and industry [34,[41][42][43][50][51][52][53][54] have shown the benefits of PnM and PuM for a wide range of workloads from different domains. However, fully adopting PIM in commercial systems is still very challenging due to the lack of tools and system support for PIM architectures across the computer architecture stack [4], which includes: (i) workload characterization methodologies and benchmark suites targeting PIM architectures; (ii) frameworks that can facilitate the implementation of complex operations and algorithms using the underlying PIM primitives (e.g., simple PIM arithmetic operations [19], bulk bitwise Boolean in-DRAM operations [83,84,92]); (iii) compiler support and compiler optimizations targeting PIM architectures; (iv) operating system support for PIM-aware virtual memory, memory management, data allocation and mapping; and (v) efficient data coherence and consistency mechanisms.…”
Section: Motivation and Problemmentioning
confidence: 99%
“…Finally, the last two bars show the impact of applying various recent technologies such as DRAM refresh reduction [39], DRAM idle power-off [56], memory disaggregation [44], and near-memory processing [38]. We model the impact of these optimizations by employing them only on the appropriate components, following the typical power profile of data-centerclass servers [26] and switches [4], [28].…”
Section: Motivationmentioning
confidence: 99%
“…To capture a wider range or projections, we also include two more sophisticated energy-efficiency optimizations that have been in active development in industry for more than a decade, and while they are not mainstream products yet, they are in advanced stages of development. These optimizations include near-memory processing [38] and disaggregation [44]. There is a large number of much more aggressive optimizations that we explicitly chose not to include, as their ability to scale up to production at reasonable cost is unknown, or they are not a good fit for hypercale data centers, or simply because they are not commercially available yet, despite their high potential (e.g., STT-RAM, PCM, near-threshold-voltage processors, spintronics, neuromorphic processors, and chipand board-level photonics).…”
Section: Motivationmentioning
confidence: 99%
“…In such a computing architecture, data can be processed directly inside the memory, minimizing the data movement between the CPU and the memory. Machine learning applications [26,29], databases [15,16], personalised recommendation systems [10,11], and genomics [2] benefit from the massive parallelization of in-memory computing.…”
Section: Introductionmentioning
confidence: 99%