Command Palette
Search for a command to run...
TOPIQ: A Top-down Approach from Semantics to Distortions for Image Quality Assessment
Chaofeng Chen Jiadi Mo Jingwen Hou Haoning Wu Liang Liao Wenxiu Sun Qiong Yan Weisi Lin

Abstract
Image Quality Assessment (IQA) is a fundamental task in computer vision that has witnessed remarkable progress with deep neural networks. Inspired by the characteristics of the human visual system, existing methods typically use a combination of global and local representations (\ie, multi-scale features) to achieve superior performance. However, most of them adopt simple linear fusion of multi-scale features, and neglect their possibly complex relationship and interaction. In contrast, humans typically first form a global impression to locate important regions and then focus on local details in those regions. We therefore propose a top-down approach that uses high-level semantics to guide the IQA network to focus on semantically important local distortion regions, named as \emph{TOPIQ}. Our approach to IQA involves the design of a heuristic coarse-to-fine network (CFANet) that leverages multi-scale features and progressively propagates multi-level semantic information to low-level representations in a top-down manner. A key component of our approach is the proposed cross-scale attention mechanism, which calculates attention maps for lower level features guided by higher level features. This mechanism emphasizes active semantic regions for low-level distortions, thereby improving performance. CFANet can be used for both Full-Reference (FR) and No-Reference (NR) IQA. We use ResNet50 as its backbone and demonstrate that CFANet achieves better or competitive performance on most public FR and NR benchmarks compared with state-of-the-art methods based on vision transformers, while being much more efficient (with only ${\sim}13\%$ FLOPS of the current best FR method). Codes are released at \url{https://github.com/chaofengc/IQA-PyTorch}.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ (IAA) | KLCC: 0.40663 PLCC: 0.51061 SROCC: 0.51687 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ trained on PIPAL | KLCC: 0.42811 PLCC: 0.57564 SROCC: 0.55568 Type: FR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ trained on SPAQ (NR) | KLCC: 0.53140 PLCC: 0.60905 SROCC: 0.64923 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ + Res50 (IAA) | KLCC: 0.28473 PLCC: 0.34000 SROCC: 0.36204 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ FACE | KLCC: 0.48428 PLCC: 0.58949 SROCC: 0.59564 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ | KLCC: 0.46217 PLCC: 0.57955 SROCC: 0.57341 Type: FR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ | KLCC: 0.50670 PLCC: 0.57674 SROCC: 0.62715 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | TOPIQ trained on FLIVE | KLCC: 0.26774 PLCC: 0.33940 SROCC: 0.34092 Type: NR |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.