Towards Conceptual Compression

We introduce a simple recurrent variational auto-encoder architecture that significantly improves image modeling. The system represents the state-of-the-art in latent variable models for both the ImageNet and Omniglot datasets. We show that it naturally separates global conceptual information from lower level details, thus addressing one of the fundamentally desired properties of unsupervised learning. Furthermore, the possibility of restricting ourselves to storing only global information about an image allows us to achieve high quality 'conceptual compression'.

PDF Abstract NeurIPS 2016 PDF NeurIPS 2016 Abstract

Datasets


Results from the Paper


Ranked #64 on Image Generation on CIFAR-10 (bits/dimension metric)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Image Generation CIFAR-10 Conv DRAW bits/dimension 3.58 # 64

Methods


No methods listed for this paper. Add relevant methods here