Discriminator Contrastive Divergence: Semi-Amortized Generative Modeling by Exploring Energy of the Discriminator

5 Apr 2020  ·  Yuxuan Song, Qiwei Ye, Minkai Xu, Tie-Yan Liu ·

Generative Adversarial Networks (GANs) have shown great promise in modeling high dimensional data. The learning objective of GANs usually minimizes some measure discrepancy, \textit{e.g.}, $f$-divergence~($f$-GANs) or Integral Probability Metric~(Wasserstein GANs). With $f$-divergence as the objective function, the discriminator essentially estimates the density ratio, and the estimated ratio proves useful in further improving the sample quality of the generator. However, how to leverage the information contained in the discriminator of Wasserstein GANs (WGAN) is less explored. In this paper, we introduce the Discriminator Contrastive Divergence, which is well motivated by the property of WGAN's discriminator and the relationship between WGAN and energy-based model. Compared to standard GANs, where the generator is directly utilized to obtain new samples, our method proposes a semi-amortized generation procedure where the samples are produced with the generator's output as an initial state. Then several steps of Langevin dynamics are conducted using the gradient of the discriminator. We demonstrate the benefits of significant improved generation on both synthetic data and several real-world image generation benchmarks.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Image Generation CIFAR-10 SNGAN-DCD (Latent) Inception score 9.11 # 26
FID 16.24 # 114
Image Generation CIFAR-10 SNGAN-DCD (Pixel) Inception score 8.54 # 37
FID 21.67 # 124
Image Generation STL-10 SNGAN-DCD (Latent) FID 17.68 # 7
Inception score 9.33 # 12
Image Generation STL-10 SNGAN-DCD (Pixel) FID 22.25 # 11
Inception score 9.25 # 13

Methods