2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.577
|View full text |Cite
|
Sign up to set email alerts
|

Full Resolution Image Compression with Recurrent Neural Networks

Abstract: This paper presents a set of full-resolution lossy image compression methods based on neural networks. Each of the architectures we describe can provide variable compression rates during deployment without requiring retraining of the network: each network need only be trained once. All of our architectures consist of a recurrent neural network (RNN)-based encoder and decoder, a binarizer, and a neural network for entropy coding. We compare RNN types (LSTM, associative LSTM) and introduce a new hybrid of GRU an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

3
802
0
1

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 814 publications
(835 citation statements)
references
References 18 publications
(24 reference statements)
3
802
0
1
Order By: Relevance
“…Like [29], soft relaxation of quantization is used to resolve the nondifferentiability problem of the quantization function [27]. Different from [27,29], the thumbnail images are compressed by a recurrent neural networks architecture, in which a stochastic rounding operation makes feature maps binarized [30]. Recently, a virtual codec network has been learned to imitate the projection from the represented vectors to the decoded images to make the image compression framework trainable in an end-to-end way [11].…”
Section: B Deep Image Compression Frameworkmentioning
confidence: 99%
“…Like [29], soft relaxation of quantization is used to resolve the nondifferentiability problem of the quantization function [27]. Different from [27,29], the thumbnail images are compressed by a recurrent neural networks architecture, in which a stochastic rounding operation makes feature maps binarized [30]. Recently, a virtual codec network has been learned to imitate the projection from the represented vectors to the decoded images to make the image compression framework trainable in an end-to-end way [11].…”
Section: B Deep Image Compression Frameworkmentioning
confidence: 99%
“…In deep image compression [1], [7], [8], the handcrafted analysis and synthesis transforms are replaced by the encoder z = f (x; θ) and decoderx = g (ẑ; φ) of a convolutional autoencoder, parametrized by θ and φ. The fundamental difference is that the transforms are not designed but learned from training data.…”
Section: Introductionmentioning
confidence: 99%
“…The problem is solved using gradient descent and backpropagation [18]. To make the model differentiable, which is required to apply backpropagation, during training the quantizer is replaced by a differentiable proxy function [1], [7], [8]. Similarly, entropy coding is invertible, but it is necessary to compute the length of the bitstream b.…”
Section: Introductionmentioning
confidence: 99%
“…For example, some image compression approaches use generative models to learn the distribution of images using adversarial training [6,7,8] to achieved impressive subjective quality at extremely low bit rate. Some works use recurrent neural networks to compress the residual information recursively, such as [9,10,11] to realize scalable coding. Some approaches propose a hyperpriorbased and context-adaptive context model to compress codes effectively in [12,13,14].…”
Section: Introductionmentioning
confidence: 99%