2019
DOI: 10.48550/arxiv.1912.10088
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

From Patches to Pictures (PaQ-2-PiQ): Mapping the Perceptual Space of Picture Quality

Abstract: Blind or no-reference (NR) perceptual picture quality prediction is a difficult, unsolved problem of great consequence to the social and streaming media industries that impacts billions of viewers daily. Unfortunately, popular NR prediction models perform poorly on real-world distorted pictures. To advance progress on this problem, we introduce the largest (by far) subjective picture quality database, containing about 40000 real-world distorted pictures and 120000 patches, on which we collected about 4M human … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 64 publications
0
3
0
Order By: Relevance
“…MLSP-FF [8] also uses heavier Inception-ResNet-V2 [33] for feature extraction. Some methods [40,41] use the features extractor pre-trained with IQA datasets [13,39]. PVQ [40] also extracts features pretrained on action recognition dataset [16] for better perception on inter-frame distortion.…”
Section: Related Workmentioning
confidence: 99%
“…MLSP-FF [8] also uses heavier Inception-ResNet-V2 [33] for feature extraction. Some methods [40,41] use the features extractor pre-trained with IQA datasets [13,39]. PVQ [40] also extracts features pretrained on action recognition dataset [16] for better perception on inter-frame distortion.…”
Section: Related Workmentioning
confidence: 99%
“…Early models were often distortion-specific, targeting only one or more specific distortions. Later, more general models have involved training quality prediction models to learn mappings from features to subjective judgments, e.g., BRISQUE [27], V-BLIINDS [28] and BIQME [29], which make use of natural scene statistics (NSS) models; BPRI [30], [31], which utilizes a pseudo reference image, TLVQM [32], which deploys a simplified motion estimator, and more recent deep learning models, like NIMA [33], PaQ-2-PiQ [34], PQR [35], DLIQA [36], and more [37]. There are also completely blind (unsupervised) models, like NIQE [38] and IL-NIQE [39].…”
Section: Related Workmentioning
confidence: 99%
“…With an interest in its possibilities, Deepti Ghadiyaram and I published one of the earliest papers, if not the first, on picture quality prediction using deep (belief) nets [5]. Unfortunately, when trained and tested on the new LIVE Challenge picture quality database, which was the largest database then available [6] (but see [7]), the deep models could not attain the performance of simple natural scene-based models like BRISQUE [8] and FRIQUEE [9]. Rigorous tests on other researchers' deep models produced similar results [10].…”
Section: Denialmentioning
confidence: 99%