Densely connected normalizing flows

Normalizing flows are bijective mappings between inputs and latent representations with a fully factorized distribution. They are very attractive due to exact likelihood valuation and efficient sampling. However, their effective capacity is often insufficient since the bijectivity constraint limits the model width. We address this issue by incrementally padding intermediate representations with noise. We precondition the noise in accordance with previous invertible units, which we describe as cross-unit coupling. Our invertible glow-like modules increase the model expressivity by fusing a densely connected block with Nystrom self-attention. We refer to our architecture as DenseFlow since both cross-unit and intra-module couplings rely on dense connectivity. Experiments show significant improvements due to the proposed contributions and reveal state-of-the-art density estimation under moderate computing budgets.

PDF Abstract NeurIPS 2021 PDF NeurIPS 2021 Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Image Generation CelebA 64x64 DenseFlow-74-10 bits/dimension 1.99 # 5
Image Generation CIFAR-10 DenseFlow-74-10 FID 34.90 # 137
bits/dimension 2.98 # 30
Image Generation ImageNet 32x32 DenseFlow-74-10 bpd 3.63 # 2
Image Generation ImageNet 64x64 DenseFlow-74-10 Bits per dim 3.35 # 2

Methods