Command Palette
Search for a command to run...
Xinyi Ying Longguang Wang Yingqian Wang Weidong Sheng Wei An Yulan Guo

Abstract
The spatio-temporal information among video sequences is significant for video super-resolution (SR). However, the spatio-temporal information cannot be fully used by existing video SR methods since spatial feature extraction and temporal motion compensation are usually performed sequentially. In this paper, we propose a deformable 3D convolution network (D3Dnet) to incorporate spatio-temporal information from both spatial and temporal dimensions for video SR. Specifically, we introduce deformable 3D convolution (D3D) to integrate deformable convolution with 3D convolution, obtaining both superior spatio-temporal modeling capability and motion-aware modeling flexibility. Extensive experiments have demonstrated the effectiveness of D3D in exploiting spatio-temporal information. Comparative results show that our network achieves state-of-the-art SR performance. Code is available at: https://github.com/XinyiYing/D3Dnet.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| video-super-resolution-on-msu-vsr-benchmark | D3Dnet | 1 - LPIPS: 0.915 ERQAv1.0: 0.674 FPS: 0.041 PSNR: 29.703 QRCRv1.0: 0.549 SSIM: 0.876 Subjective score: 5.066 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.