2018
DOI: 10.48550/arxiv.1806.00681
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Nonlocal Neural Networks, Nonlocal Diffusion and Nonlocal Modeling

Yunzhe Tao,
Qi Sun,
Qiang Du
et al.

Abstract: Nonlocal neural networks [25] have been proposed and shown to be effective in several computer vision tasks, where the nonlocal operations can directly capture long-range dependencies in the feature space. In this paper, we study the nature of diffusion and damping effect of nonlocal networks by doing spectrum analysis on the weight matrices of the well-trained networks, and then propose a new formulation of the nonlocal block. The new block not only learns the nonlocal interactions but also has stable dynamic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 17 publications
(32 reference statements)
0
5
0
Order By: Relevance
“…The Nonlocal block [31] can capture remote dependencies more robustly and flexibly to help the deep network better integrate Nonlocal information. Some Nonlocal network modules currently proposed are NL [31], A2 [32], NS [33], CC [34], CGNL [35], SNL [36], etc. Chen et al [32] propose the Double Attention block, which first collects the features in the entire space and then assigns them back to each location.…”
Section: A Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The Nonlocal block [31] can capture remote dependencies more robustly and flexibly to help the deep network better integrate Nonlocal information. Some Nonlocal network modules currently proposed are NL [31], A2 [32], NS [33], CC [34], CGNL [35], SNL [36], etc. Chen et al [32] propose the Double Attention block, which first collects the features in the entire space and then assigns them back to each location.…”
Section: A Related Workmentioning
confidence: 99%
“…Huang et al [34] proposed a lightweight Nonlocal block called an interleaved attention block, which decomposes the positional attention of NL into conterminously column-wise and row-wise attention. In order to improve the stability of the NL block, Tao et al [33] proposed using the Laplacian of the incidence matrix as an attention map, and the Nonlocal stage (NS) module can follow the diffusion characteristics. Zhu et al [36] proposed The SNL (Spectral Nonlocal Block) module, which symmetrically processes the attention feature block.…”
Section: A Related Workmentioning
confidence: 99%
“…If f q(l−1) j incorporates complementary information or more significant cues compared to the current probe feature f q(l−1) i , then our redundency-aware attention scheme will eliminate the information from the inferior f q(l−1) i and emphasis the more discriminative f q(l−1) j . Compared to the original non-local network which uses only f q(l−1) j [19], our formulation can be more similar to the diffusion maps [58], graph Laplacian [59] and non-local image processing [60]. All of them are non-local analogues [61] of local diffusions, which are expected to be more stable than its original non-local counterpart [19] due to the nature of its inherit Hilbert-Schmidt operator [61].…”
Section: Redundancy-aware Self-attentionmentioning
confidence: 99%
“…If x l−1 n j incorporates complementary information and has better imaging/content quality compared to x l−1 n i , then RSA will erase some information of the inferior x l−1 n i and replaces it by the more discriminative feature representation x l−1 n j . Compared to the method of using only x l−1 n j [75], our setting shares more common features with diffusion maps [70], graph Laplacian [9] and non-local image processing [17]. All of them are non-local analogues [12] of local diffusions, which are expected to be more stable than its original non-local counterpart [75] due to the nature of its inherit Hilbert-Schmidt operator [12].…”
Section: Global 3d Encoding With Arbitrary Number Of Appearance Descr...mentioning
confidence: 99%