2020
DOI: 10.1109/tip.2020.2977457
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Objective Matrix Normalization for Fine-Grained Visual Recognition

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
4
1

Relationship

2
8

Authors

Journals

citations
Cited by 66 publications
(22 citation statements)
references
References 64 publications
0
20
0
Order By: Relevance
“…They evaluated their model using realistic datasets such as HMDB51 [23] and UCF101 and showed that the HC-MTL method outperformed other methods for both action grouping and recognition. In addition, Min et al [24] proposed a Multi-Objective Matrix Normalization (MOMN) method for fine-grained visual recognition. Their proposed system can simultaneously normalize a bilinear representation in square-root, low-rank, and sparsity.…”
Section: Literature Reviewmentioning
confidence: 99%
“…They evaluated their model using realistic datasets such as HMDB51 [23] and UCF101 and showed that the HC-MTL method outperformed other methods for both action grouping and recognition. In addition, Min et al [24] proposed a Multi-Objective Matrix Normalization (MOMN) method for fine-grained visual recognition. Their proposed system can simultaneously normalize a bilinear representation in square-root, low-rank, and sparsity.…”
Section: Literature Reviewmentioning
confidence: 99%
“…iSQRT-COV [101] and the improved B-CNN [146] used the Newton-Schulz iteration to approximate matrix square-root normalization with only matrix multiplication to decrease training time. Recently, MOMN [106] was proposed to simultaneously normalize a bilinear representation in terms of square-root, low-rank, and sparsity all within a multi-objective optimization framework.…”
Section: Performing High-order Feature Interactionsmentioning
confidence: 99%
“…All the neural network models are implemented using the PyTorch framework. We compare various methods, including deep networks (e.g., DenseNet161 [55] and SENet [56]), recently proposed fine-grained recognition methods (e.g., MOMN [57] and PMG [58]) and food recognition methods (e.g., PAR-Net [9]). For the deep networks, we train all the networks with parameters initialized from ImageNet pretrained weights with a learning rate of 10 −2 , and divided by 10 after 30 epoches.…”
Section: Implementation Detailsmentioning
confidence: 99%