Command Palette
Search for a command to run...
Jianyi Wang Kelvin C.K. Chan Chen Change Loy

Abstract
Measuring the perception of visual content is a long-standing problem in computer vision. Many mathematical models have been developed to evaluate the look or quality of an image. Despite the effectiveness of such tools in quantifying degradations such as noise and blurriness levels, such quantification is loosely coupled with human language. When it comes to more abstract perception about the feel of visual content, existing methods can only rely on supervised models that are explicitly trained with labeled data collected via laborious user study. In this paper, we go beyond the conventional paradigms by exploring the rich visual language prior encapsulated in Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner. In particular, we discuss effective prompt designs and show an effective prompt pairing strategy to harness the prior. We also provide extensive experiments on controlled datasets and Image Quality Assessment (IQA) benchmarks. Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments. Code is avaliable at https://github.com/IceClear/CLIP-IQA.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| no-reference-image-quality-assessment-on-uhd | CLIP-IQA+ | PLCC: 0.709 SRCC: 0.747 |
| video-quality-assessment-on-msu-sr-qa-dataset | ClipIQA+ | KLCC: 0.69774 PLCC: 0.71808 SROCC: 0.56875 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | ClipIQA+ ViT-L-14 | KLCC: 0.38794 PLCC: 0.50379 SROCC: 0.49881 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | ClipIQA | KLCC: 0.49417 PLCC: 0.58944 SROCC: 0.60808 Type: NR |
| video-quality-assessment-on-msu-sr-qa-dataset | ClipIQA+ ResNet50 | KLCC: 0.52628 PLCC: 0.65154 SROCC: 0.65713 Type: NR |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.