2021
DOI: 10.1109/tip.2021.3101395
|View full text |Cite
|
Sign up to set email alerts
|

WaveCNet: Wavelet Integrated CNNs to Suppress Aliasing Effect for Noise-Robust Image Classification

Abstract: Though widely used in image classification, convolutional neural networks (CNNs) are prone to noise interruptions, i.e. the CNN output can be drastically changed by small image noise. To improve the noise robustness, we try to integrate CNNs with wavelet by replacing the common down-sampling (maxpooling, strided-convolution, and average pooling) with discrete wavelet transform (DWT). We firstly propose general DWT and inverse DWT (IDWT) layers applicable to various orthogonal and biorthogonal discrete wavelets… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
25
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 58 publications
(25 citation statements)
references
References 57 publications
(98 reference statements)
0
25
0
Order By: Relevance
“…The pooling layer generally follows the convolution layer to reduce the width and height of the feature map but does not alter the depth of the feature map. Common pooling layers include average pooling and max pooling layers 48 Fully connected layer …”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The pooling layer generally follows the convolution layer to reduce the width and height of the feature map but does not alter the depth of the feature map. Common pooling layers include average pooling and max pooling layers 48 Fully connected layer …”
Section: Methodsmentioning
confidence: 99%
“…Common pooling layers include average pooling and max pooling layers. 48 • Fully connected layer "Fully connected" means that the output of the previous layer is distinctly connected to individual neurons in the next layer. 49 The aim of the fully connected layer is to use the high-level feature of the input image produced by convolutional and pooling layers to classify the input image based on the training dataset.…”
Section: • Pooling Layermentioning
confidence: 99%
“…They modified and combined the WLFD by generating wavelet pyramids, keypoint localization, and descriptors. Another study in [25] combined a wavelet with the convolutional neural network (WaveCNet) to produce better noise-robustness. In [26], a scale and rotation invariant wavelet feature transform was proposed using a biorthogonal wavelet and combined only two sub-bands with SIFT.…”
Section: A Literature Reviewmentioning
confidence: 99%
“…The set of all unit vectors in the tangent plane of S is a circle, hence a function of κ in (25). The maximum and minimum values of κ 1 and κ 2 of κ are the principal curvatures of surface S at p. The mean curvature H (26) and Gaussian curvature K (27) can be calculated from the principal curvature of κ 1 and κ 2 [45].…”
Section: B Surface Curvaturementioning
confidence: 99%
“…Chang et al 9 added depth information to the loss function and used the change of depth value on the edge of the classified object to constrain the network training. However, there are still many problems with the above approach: Factors effecting on the one hand, the indoor noise exists in different frequency part of image more 10 , while traditional convolution neural network down sampling operations such as average pooling, maximum pooling will not separate different frequency information, which can lead to high frequency noise increased with the increase of the depth of the network are preserved, the sampling data of aliasing in the basic structure of the residual noise destroys the image features, Thus, it brings difficulties to the image segmentation task. On the other hand, the depth image is used as the fourth channel to fuse with the color image, which does not make full use of the complementarity of RGB color information and depth information.…”
Section: Introductionmentioning
confidence: 99%