2016
DOI: 10.1007/s11220-016-0152-5
|View full text |Cite
|
Sign up to set email alerts
|

Recognizable or Not: Towards Image Semantic Quality Assessment for Compression

Abstract: Traditionally, image compression was optimized for the pixel-wise fidelity or the perceptual quality of the compressed images given a bit-rate budget. But recently, compressed images are more and more utilized for automatic semantic analysis tasks such as recognition and retrieval. For these tasks, we argue that the optimization target of compression is no longer perceptual quality, but the utility of the compressed images in the given automatic semantic analysis task. Accordingly, we propose to evaluate the q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
8
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 25 publications
(8 citation statements)
references
References 32 publications
0
8
0
Order By: Relevance
“…Moreover, there are coding schemes that serve for automatic semantic analysis instead of human viewing, such as surveillance video coding. For these schemes, the quality metric shall be semantic quality [74], which remains largely unexplored. As a special note, we find that there is a tradeoff between signal fidelity, perceptual naturalness, and semantic quality [75], which implies that the optimization target shall be aligned with the actual requirement.…”
Section: Perspectives and Conclusionmentioning
confidence: 99%
“…Moreover, there are coding schemes that serve for automatic semantic analysis instead of human viewing, such as surveillance video coding. For these schemes, the quality metric shall be semantic quality [74], which remains largely unexplored. As a special note, we find that there is a tradeoff between signal fidelity, perceptual naturalness, and semantic quality [75], which implies that the optimization target shall be aligned with the actual requirement.…”
Section: Perspectives and Conclusionmentioning
confidence: 99%
“…1) Full-Reference Perceptual Loss Comparison on SOTS: Since dehazed images are often subsequently fed for automatic semantic analysis tasks such as recognition and detection, we argue that the optimization target of dehazing in these tasks is neither pixel-level or perceptual-level quality, but the utility of the dehazed images in the given semantic analysis task [59]. The perceptual loss [32] was proposed to measure the semantic-level similarity of images, using the VGG recognition model 4 pre-trained on ImageNet dataset [60].…”
Section: B Restoration Versus High-level Visionmentioning
confidence: 99%
“…While it's a contradictory that most complicated quality metrics with high performance are not able to be integrated easily into an image compression loop. Some research works tried to do this by adjusting image compression parameters (e.g., Quantization parameters) heuristically according to embedded quality metrics [5,6], but they are still not fully automatic-optimized end-to-end image encoder with integration of complicated distortion metrics.…”
Section: Introductionmentioning
confidence: 99%
“…https://github.com/tensorflow/models/tree/master/research/compression6 A fixed header size of 100 bytes in JPEG2000 is added for all results.…”
mentioning
confidence: 99%