2023
DOI: 10.1109/tifs.2023.3262149
|View full text |Cite
|
Sign up to set email alerts
|

FastSecNet: An Efficient Cryptographic Framework for Private Neural Network Inference

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 49 publications
0
7
0
Order By: Relevance
“…A recent art Cheetah [13] proposes a special encoding method to encode vectors and matrices into HE polynomials, which achieves state-of-the-art performance in computing matrix-vector multiplication and convolutions. Iron [10] realizes that matrix-matrix multiplication (rather than matrix-vector multiplication) dominates in transformer-based inference, and therefore improves the vanilla polynomial encoding by introducing a blocking method that prioritizes the batch dimension. Despite the optimization, some of the non-linear functions (e.g., GELU, softmax and layer normalization layers) are fundamentally expensive in private inference.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…A recent art Cheetah [13] proposes a special encoding method to encode vectors and matrices into HE polynomials, which achieves state-of-the-art performance in computing matrix-vector multiplication and convolutions. Iron [10] realizes that matrix-matrix multiplication (rather than matrix-vector multiplication) dominates in transformer-based inference, and therefore improves the vanilla polynomial encoding by introducing a blocking method that prioritizes the batch dimension. Despite the optimization, some of the non-linear functions (e.g., GELU, softmax and layer normalization layers) are fundamentally expensive in private inference.…”
Section: Related Workmentioning
confidence: 99%
“…Despite the optimization, some of the non-linear functions (e.g., GELU, softmax and layer normalization layers) are fundamentally expensive in private inference. For instance, Iron [10] reports that running a single inference on BERT-Tiny [3] requires 50 seconds time and 2GB transmission. Two recent studies explore replacing these fundamentally expensive non-linear functions with operators that are more friendly in private inference.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations