2023
DOI: 10.21203/rs.3.rs-3546552/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SAFMem: Accelerating Transformer Self-Attention Functionality via Memristor-Based Hardware

Meriem Bettayeb,
Yasmin Halawani,
Muhammad Umair Khan
et al.

Abstract: The adoption of transformer networks has experienced a notable surge in various AI applications. However, the increasedcomputational complexity, stemming primarily from the self-attention mechanism, parallels the manner in which convolutionoperations constrain the capabilities and speed of Convolutional Neural Networks (CNNs). The self-attention algorithm,specifically the Matrix-matrix Multiplication (MatMul) operations, demands a substantial amount of memory and computationalcomplexity, thereby restricting th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
references
References 39 publications
(43 reference statements)
0
0
0
Order By: Relevance