Command Palette
Search for a command to run...
Daniel Seichter Söhnke Benedikt Fischedick Mona Köhler Horst-Michael Groß

Abstract
Semantic scene understanding is essential for mobile agents acting in various environments. Although semantic segmentation already provides a lot of information, details about individual objects as well as the general scene are missing but required for many real-world applications. However, solving multiple tasks separately is expensive and cannot be accomplished in real time given limited computing and battery capabilities on a mobile platform. In this paper, we propose an efficient multi-task approach for RGB-D scene analysis~(EMSANet) that simultaneously performs semantic and instance segmentation~(panoptic segmentation), instance orientation estimation, and scene classification. We show that all tasks can be accomplished using a single neural network in real time on a mobile platform without diminishing performance - by contrast, the individual tasks are able to benefit from each other. In order to evaluate our multi-task approach, we extend the annotations of the common RGB-D indoor datasets NYUv2 and SUNRGB-D for instance segmentation and orientation estimation. To the best of our knowledge, we are the first to provide results in such a comprehensive multi-task setting for indoor scene analysis on NYUv2 and SUNRGB-D.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| panoptic-segmentation-on-nyu-depth-v2 | EMSANet | PQ: 47.38 |
| panoptic-segmentation-on-sun-rgbd | EMSANet | PQ: 52.84 |
| semantic-segmentation-on-nyu-depth-v2 | EMSANet (2x ResNet-34 NBt1D, finetuned) | Mean IoU: 53.34% |
| semantic-segmentation-on-sun-rgbd | DPLNet | Mean IoU: 48.47% |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.