Command Palette
Search for a command to run...
Taesung Park; Ming-Yu Liu; Ting-Chun Wang; Jun-Yan Zhu

Abstract
We propose spatially-adaptive normalization, a simple but effective layer for synthesizing photorealistic images given an input semantic layout. Previous methods directly feed the semantic layout as input to the deep network, which is then processed through stacks of convolution, normalization, and nonlinearity layers. We show that this is suboptimal as the normalization layers tend to ``wash away'' semantic information. To address the issue, we propose using the input layout for modulating the activations in normalization layers through a spatially-adaptive, learned transformation. Experiments on several challenging datasets demonstrate the advantage of the proposed method over existing approaches, regarding both visual fidelity and alignment with input layouts. Finally, our model allows user control over both semantic and style. Code is available at https://github.com/NVlabs/SPADE .
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| image-to-image-translation-on-ade20k-labels | SPADE | Accuracy: 79.9% FID: 33.9 LPIPS: 0 mIoU: 38.5 |
| image-to-image-translation-on-ade20k-outdoor | SPADE | Accuracy: 82.9% FID: 63.3 mIoU: 30.8 |
| image-to-image-translation-on-cityscapes | SPADE | FID: 71.8 Per-pixel Accuracy: 81.9% mIoU: 62.3 |
| image-to-image-translation-on-coco-stuff | SPADE | Accuracy: 67.9% FID: 22.6 mIoU: 37.4 |
| sketch-to-image-translation-on-coco-stuff | SPADE | FID: 89.2 FID-C: 48.9 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.