Command Palette
Search for a command to run...
Shariq Farooq Bhat Ibraheem Alhashim Peter Wonka

Abstract
We address the problem of estimating a high quality dense depth map from a single RGB input image. We start out with a baseline encoder-decoder convolutional neural network architecture and pose the question of how the global processing of information can help improve overall depth estimation. To this end, we propose a transformer-based architecture block that divides the depth range into bins whose center value is estimated adaptively per image. The final depth values are estimated as linear combinations of the bin centers. We call our new building block AdaBins. Our results show a decisive improvement over the state-of-the-art on several popular depth datasets across all metrics. We also validate the effectiveness of the proposed block with an ablation study and provide the code and corresponding pre-trained weights of the new state-of-the-art model.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| depth-estimation-on-nyu-depth-v2 | AdaBins | RMS: 0.364 |
| monocular-depth-estimation-on-kitti-eigen | AdaBins | Delta u003c 1.25: 0.964 Delta u003c 1.25^2: 0.995 Delta u003c 1.25^3: 0.999 RMSE: 2.360 RMSE log: 0.088 absolute relative error: 0.058 |
| monocular-depth-estimation-on-nyu-depth-v2 | AdaBins | Delta u003c 1.25: 0.903 Delta u003c 1.25^2: 0.984 Delta u003c 1.25^3: 0.997 RMSE: 0.364 absolute relative error: 0.103 log 10: 0.044 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.