Command Palette
Search for a command to run...
Gaurav Parmar Dacheng Li Kwonjoon Lee Zhuowen Tu

Abstract
We present a new generative autoencoder model with dual contradistinctive losses to improve generative autoencoder that performs simultaneous inference (reconstruction) and synthesis (sampling). Our model, named dual contradistinctive generative autoencoder (DC-VAE), integrates an instance-level discriminative loss (maintaining the instance-level fidelity for the reconstruction/synthesis) with a set-level adversarial loss (encouraging the set-level fidelity for there construction/synthesis), both being contradistinctive. Extensive experimental results by DC-VAE across different resolutions including 32x32, 64x64, 128x128, and 512x512 are reported. The two contradistinctive losses in VAE work harmoniously in DC-VAE leading to a significant qualitative and quantitative performance enhancement over the baseline VAEs without architectural changes. State-of-the-art or competitive results among generative autoencoders for image reconstruction, image synthesis, image interpolation, and representation learning are observed. DC-VAE is a general-purpose VAE model, applicable to a wide variety of downstream tasks in computer vision and machine learning.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| image-generation-on-celeba-128x128 | DC-VAE | FID: 19.9 |
| image-generation-on-celeba-hq-256x256 | DC-VAE | FID: 15.81 |
| image-generation-on-cifar-10 | DC-VAE | FID: 17.9 |
| image-generation-on-lsun-bedroom-128-x-128 | DC-VAE | FID: 14.3 |
| image-generation-on-stl-10 | DC-VAE | FID: 41.9 Inception score: 8.1 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.