Command Palette
Search for a command to run...
T2Net: Synthetic-to-Realistic Translation for Solving Single-Image Depth Estimation Tasks
Chuanxia Zheng; Tat-Jen Cham; Jianfei Cai

Abstract
Current methods for single-image depth estimation use training datasets with real image-depth pairs or stereo pairs, which are not easy to acquire. We propose a framework, trained on synthetic image-depth pairs and unpaired real images, that comprises an image translation network for enhancing realism of input images, followed by a depth prediction network. A key idea is having the first network act as a wide-spectrum input translator, taking in either synthetic or real images, and ideally producing minimally modified realistic images. This is done via a reconstruction loss when the training input is real, and GAN loss when synthetic, removing the need for heuristic self-regularization. The second network is trained on a task loss for synthetic image-depth pairs, with extra GAN loss to unify real and synthetic feature distributions. Importantly, the framework can be trained end-to-end, leading to good results, even surpassing early deep-learning methods that use real paired data.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| depth-estimation-on-dcm | T2Net | Abs Rel: 0.351 RMSE: 1.117 RMSE log: 0.415 Sq Rel: 0.416 |
| depth-estimation-on-ebdtheque | T2Net | Abs Rel: 0.491 RMSE: 1.459 RMSE log: 0.777 Sq Rel: 0.555 |
| unsupervised-domain-adaptation-on-virtual-2 | T2Net | RMSE : 4.674 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.