1 code implementation • 16 Aug 2024 • Julius Kunze, Daniel Severo, Giulio Zani, Jan-Willem van de Meent, James Townsend
We present shuffle coding, a general method for optimal compression of sequences of unordered objects using bits-back coding.
2 code implementations • 5 Oct 2022 • Ruben Villegas, Mohammad Babaeizadeh, Pieter-Jan Kindermans, Hernan Moraldo, Han Zhang, Mohammad Taghi Saffar, Santiago Castro, Julius Kunze, Dumitru Erhan
To the best of our knowledge, this is the first time a paper studies generating videos from time variable prompts.
Ranked #4 on Video Prediction on BAIR Robot Pushing
1 code implementation • 30 Nov 2021 • Julius Kunze, James Townsend, David Barber
We propose a new, more general approach to the design of stochastic gradient-based optimization methods for machine learning.
1 code implementation • ICLR 2020 • James Townsend, Thomas Bird, Julius Kunze, David Barber
We make the following striking observation: fully convolutional VAE models trained on 32x32 ImageNet can generalize well, not just to 64x64 but also to far larger photographs, with no changes to the model.
no code implementations • 12 Feb 2019 • Julius Kunze, Louis Kirsch, Hippolyt Ritter, David Barber
Variational inference with a factorized Gaussian posterior estimate is a widely used approach for learning parameters and hidden variables.
no code implementations • NeurIPS 2018 • Louis Kirsch, Julius Kunze, David Barber
Scaling model capacity has been vital in the success of deep learning.
no code implementations • 27 Sep 2018 • Julius Kunze, Louis Kirsch, Hippolyt Ritter, David Barber
We propose Noisy Information Bottlenecks (NIB) to limit mutual information between learned parameters and the data through noise.
no code implementations • 13 Sep 2018 • Thomas Bird, Julius Kunze, David Barber
These approaches are of particular interest because they are parallelizable.
3 code implementations • WS 2017 • Julius Kunze, Louis Kirsch, Ilia Kurenkov, Andreas Krug, Jens Johannsmeier, Sebastian Stober
End-to-end training of automated speech recognition (ASR) systems requires massive data and compute resources.