Command Palette
Search for a command to run...
Learning a No-Reference Quality Metric for Single-Image Super-Resolution
Chao Ma; Chih-Yuan Yang; Xiaokang Yang; Ming-Hsuan Yang

Abstract
Numerous single-image super-resolution algorithms have been proposed in the literature, but few studies address the problem of performance evaluation based on visual perception. While most super-resolution images are evaluated by fullreference metrics, the effectiveness is not clear and the required ground-truth images are not always available in practice. To address these problems, we conduct human subject studies using a large set of super-resolution images and propose a no-reference metric learned from visual perceptual scores. Specifically, we design three types of low-level statistical features in both spatial and frequency domains to quantify super-resolved artifacts, and learn a two-stage regression model to predict the quality scores of super-resolution images without referring to ground-truth images. Extensive experimental results show that the proposed metric is effective and efficient to assess the quality of super-resolution images based on human perception.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| video-quality-assessment-on-msu-sr-qa-dataset | Ma-Metric | KLCC: 0.52301 PLCC: 0.65357 SROCC: 0.67362 Type: NR |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.