2001
DOI: 10.1007/3-540-44668-0_30
|View full text |Cite
|
Sign up to set email alerts
|

Weight Quantization for Multi-layer Perceptrons Using Soft Weight Sharing

Abstract: Abstract. We propose a novel approach for quantizing the weights of a multi-layer perceptron (MLP) for efficient VLSI implementation. Our approach uses soft weight sharing, previously proposed for improved generalization and considers the weights not as constant numbers but as random variables drawn from a Gaussian mixture distribution; which includes as its special cases k-means clustering and uniform quantization. This approach couples the training of weights for reduced error with their quantization. Simula… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2007
2007
2018
2018

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 9 publications
(9 reference statements)
0
2
0
Order By: Relevance
“…The studies in [9,10,11] discretize the weights of a neural network according to the weights' ranges. The methods in [12] and [13] use uniform scalar parameter quantization to implement fixed-point versions of the networks.…”
Section: Introductionmentioning
confidence: 99%
“…The studies in [9,10,11] discretize the weights of a neural network according to the weights' ranges. The methods in [12] and [13] use uniform scalar parameter quantization to implement fixed-point versions of the networks.…”
Section: Introductionmentioning
confidence: 99%
“…For fixed-point implementations of DNNs, parameter quantization is also required. The studies in [19], [20] discretized the weights of a neural network according to the range of the weights. The methods in [21] and [22] used uniform scalar parameter quantization to implement fixed-point versions of the networks.…”
Section: Introductionmentioning
confidence: 99%