The platform will undergo maintenance on Sep 14 at about 9:30 AM EST and will be unavailable for approximately 1 hour.
2022 26th International Conference on Pattern Recognition (ICPR) 2022
DOI: 10.1109/icpr56361.2022.9956533
|View full text |Cite
|
Sign up to set email alerts
|

Fast matrix multiplication for binary and ternary CNNs on ARM CPU

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 22 publications
0
5
0
Order By: Relevance
“…These instructions can noticeably speed up eightbit QNN inference [16]. Fast implementations are also available for ternary [17][18][19] and binary networks [18,20]. However, binary and ternary networks still suffer from accuracy loss compared to full-precision or eight-bit quantized networks with a similar number of parameters and architecture, which limits their suitability for certain tasks.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…These instructions can noticeably speed up eightbit QNN inference [16]. Fast implementations are also available for ternary [17][18][19] and binary networks [18,20]. However, binary and ternary networks still suffer from accuracy loss compared to full-precision or eight-bit quantized networks with a similar number of parameters and architecture, which limits their suitability for certain tasks.…”
Section: Related Workmentioning
confidence: 99%
“…We also implemented floating-point, eight-bit, and four-bit matrix multiplications as suggested in [18]. The eight-bit multiplication uses gemmlowp-like [12] microkernels.…”
Section: Hardware and Softwarementioning
confidence: 99%
See 3 more Smart Citations