Proceedings of the 28th ACM International Conference on Multimedia 2020
DOI: 10.1145/3394171.3413804
|View full text |Cite
|
Sign up to set email alerts
|

Norm-in-Norm Loss with Faster Convergence and Better Performance for Image Quality Assessment

Abstract: Currently, most image quality assessment (IQA) models are supervised by the MAE or MSE loss with empirically slow convergence. It is well-known that normalization can facilitate fast convergence. Therefore, we explore normalization in the design of loss functions for IQA. Specifically, we first normalize the predicted quality scores and the corresponding subjective quality scores. Then, the loss is defined based on the norm of the differences between these normalized values. The resulting "Norm-in-Norm" loss e… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
33
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 62 publications
(35 citation statements)
references
References 29 publications
0
33
0
Order By: Relevance
“…whereQ (d) d,i . Note that PLCC-induced loss is also considered in Ma et al (2018), Liu et al (2018) and Li et al (2020).…”
Section: Linearity-induced Lossmentioning
confidence: 99%
“…whereQ (d) d,i . Note that PLCC-induced loss is also considered in Ma et al (2018), Liu et al (2018) and Li et al (2020).…”
Section: Linearity-induced Lossmentioning
confidence: 99%
“…For NR-IQA, CNN-based methods (Bosse et al 2017;Wu et al 2020;Su et al 2020) have significantly outperformed handcrafted statistic-based approaches (Xu et al 2016) by directly extracting discriminative features from LQ images. Due to distortion diversity and content changes, the recent trend of NR-IQA (Li, Jiang, and Jiang 2020) is to involve semantic prior information by using pretrained models on classification databases, i.e., ImageNet (Deng et al 2009). And Su et al (Su et al 2020) Note that the FR-teacher is pretrained and fixed only for distillation and the trained NAR-student is applied for testing.…”
Section: Related Workmentioning
confidence: 99%
“…Three different fitness functions are considered for regression, namely the smooth-L1, the norm-in-norm [ 30 ], and the ranking hinge loss. The smooth-L1 loss is widely used for regression tasks because of its robustness to outliers.…”
Section: Facial Image Aesthetic Estimationmentioning
confidence: 99%
“…The recent norm-in-norm loss [ 30 ] facilitates faster convergence for training a CNN based (Image Quality Assessment) IQA model and also leads to better prediction performance than the mean absolute error (MAE) and mean squared error (MSE) losses. Its estimation is based on three steps: the computation of statistics, normalization based on the statistics, and loss as the norm of the differences between normalized values.…”
Section: Facial Image Aesthetic Estimationmentioning
confidence: 99%
See 1 more Smart Citation