2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00867
|View full text |Cite
|
Sign up to set email alerts
|

Towards Visually Explaining Variational Autoencoders

Abstract: Recent advances in Convolutional Neural Network (CNN) model interpretability have led to impressive progress in visualizing and understanding model predictions. In particular, gradient-based visual attention methods have driven much recent effort in using visual attention maps as a means for visual explanations. A key problem, however, is these methods are designed for classification and categorization tasks, and their extension to explaining generative models, e.g., variational autoencoders (VAE) is not trivi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
108
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 160 publications
(110 citation statements)
references
References 33 publications
2
108
0
Order By: Relevance
“…Consequently, previous approaches have attempted to enhance the models for a better reconstruction performance. For example, well-known models were combined [1], cross-channel interactions were enabled [3], a deep convolutional generative adversarial network was used [4], a de-noising technique was proposed [5], pixel-level method and patch-level method were mixed [8], an attention technique was utilized [10], [13], and encoder-decoderencoder sub-networks were employed [15]. In [9], the reconstruction performance is improved by iteratively updating the input image using a gradient descent.…”
Section: Related Work a Unsupervised Anomaly Localizationmentioning
confidence: 99%
See 1 more Smart Citation
“…Consequently, previous approaches have attempted to enhance the models for a better reconstruction performance. For example, well-known models were combined [1], cross-channel interactions were enabled [3], a deep convolutional generative adversarial network was used [4], a de-noising technique was proposed [5], pixel-level method and patch-level method were mixed [8], an attention technique was utilized [10], [13], and encoder-decoderencoder sub-networks were employed [15]. In [9], the reconstruction performance is improved by iteratively updating the input image using a gradient descent.…”
Section: Related Work a Unsupervised Anomaly Localizationmentioning
confidence: 99%
“…This method aims to generate a reference image by transforming abnormal patterns of a test image, if any, into normal patterns observed in normal training images. Efforts have been made to develop effective models and techniques for achieving better reference images [1], [3], [5], [10], [13].…”
Section: Introductionmentioning
confidence: 99%
“…This method classifies an abnormal sample using the difference in a loss function or mutual information between normal sample and abnormal sample. The autoencoder-based method is an unsupervised-based method composed of an encoder and a decoder [15] - [22]. The autoencoder-based anomaly detection method uses the image difference between input and output images [16], [18], [20], a latent space-based score [15], [17], a loss-based score [19], and the method using the autoencoder-based GAN structure [20].…”
Section: Introductionmentioning
confidence: 99%
“…We achieve state-of-the-art textural anomaly segmentation results using a relatively simple architecture. In contrast, other methods often employ more complex architectures such as the Generative Adversarial Network (GAN) [6], [7], [13], or the Variational Autoencoder (VAE) [8], [10]. In addition, our proposed method is computationally light-weight and stable during training.…”
Section: Introductionmentioning
confidence: 99%
“…Masci et al [15] adapt this architecture by using convolutional layers. These convolutional autoencoders are better suited to operating on image data and are often applied to anomaly detection in images and videos [5], [8], [10], [16].…”
Section: Introductionmentioning
confidence: 99%