Command Palette
Search for a command to run...
Eskandar George ; Abdelsamad Mohamed ; Armanious Karim ; Zhang Shuai ; Yang Bin

Abstract
Semantic Image Synthesis (SIS) is a subclass of image-to-image translationwhere a semantic layout is used to generate a photorealistic image.State-of-the-art conditional Generative Adversarial Networks (GANs) need a hugeamount of paired data to accomplish this task while generic unpairedimage-to-image translation frameworks underperform in comparison, because theycolor-code semantic layouts and learn correspondences in appearance instead ofsemantic content. Starting from the assumption that a high quality generatedimage should be segmented back to its semantic layout, we propose a newUnsupervised paradigm for SIS (USIS) that makes use of a self-supervisedsegmentation loss and whole image wavelet based discrimination. Furthermore, inorder to match the high-frequency distribution of real images, a novelgenerator architecture in the wavelet domain is proposed. We test ourmethodology on 3 challenging datasets and demonstrate its ability to bridge theperformance gap between paired and unpaired models.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| image-to-image-translation-on-ade20k-labels | USIS-Wavelet | FID: 34.5 mIoU: 16.95 |
| image-to-image-translation-on-cityscapes | USIS-Wavelet | FID: 50.14 mIoU: 42.32 |
| image-to-image-translation-on-coco-stuff | USIS-Wavelet | FID: 28.6 mIoU: 13.4 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.