Search Results for author: Ali Razavi

Found 13 papers, 6 papers with code

Aspects of scaling and scalability for flow-based sampling of lattice QCD

no code implementations14 Nov 2022 Ryan Abbott, Michael S. Albergo, Aleksandar Botev, Denis Boyda, Kyle Cranmer, Daniel C. Hackett, Alexander G. D. G. Matthews, Sébastien Racanière, Ali Razavi, Danilo J. Rezende, Fernando Romero-López, Phiala E. Shanahan, Julian M. Urban

Recent applications of machine-learned normalizing flows to sampling in lattice field theory suggest that such methods may be able to mitigate critical slowing down and topological freezing.

Vector Quantized Models for Planning

no code implementations8 Jun 2021 Sherjil Ozair, Yazhe Li, Ali Razavi, Ioannis Antonoglou, Aäron van den Oord, Oriol Vinyals

Our key insight is to use discrete autoencoders to capture the multiple possible effects of an action in a stochastic environment.

Predicting Video with VQVAE

1 code implementation2 Mar 2021 Jacob Walker, Ali Razavi, Aäron van den Oord

In recent years, the task of video prediction-forecasting future video given past video frames-has attracted attention in the research community.

Video Generation Video Prediction

Do Transformers Need Deep Long-Range Memory

1 code implementation7 Jul 2020 Jack W. Rae, Ali Razavi

Deep attention models have advanced the modelling of sequential data across many domains.

Deep Attention Language Modelling

Do Transformers Need Deep Long-Range Memory?

no code implementations ACL 2020 Jack Rae, Ali Razavi

Deep attention models have advanced the modelling of sequential data across many domains.

Deep Attention Language Modelling

Preventing Posterior Collapse with delta-VAEs

no code implementations ICLR 2019 Ali Razavi, Aäron van den Oord, Ben Poole, Oriol Vinyals

Due to the phenomenon of "posterior collapse," current latent variable generative models pose a challenging design choice that either weakens the capacity of the decoder or requires augmenting the objective so it does not only maximize the likelihood of the data.

Ranked #7 on Image Generation on ImageNet 32x32 (bpd metric)

Image Generation Representation Learning

Hyperbolic Attention Networks

no code implementations ICLR 2019 Caglar Gulcehre, Misha Denil, Mateusz Malinowski, Ali Razavi, Razvan Pascanu, Karl Moritz Hermann, Peter Battaglia, Victor Bapst, David Raposo, Adam Santoro, Nando de Freitas

We introduce hyperbolic attention networks to endow neural networks with enough capacity to match the complexity of data with hierarchical and power-law structure.

Machine Translation Question Answering +2

Population Based Training of Neural Networks

9 code implementations27 Nov 2017 Max Jaderberg, Valentin Dalibard, Simon Osindero, Wojciech M. Czarnecki, Jeff Donahue, Ali Razavi, Oriol Vinyals, Tim Green, Iain Dunning, Karen Simonyan, Chrisantha Fernando, Koray Kavukcuoglu

Neural networks dominate the modern machine learning landscape, but their training and success still suffer from sensitivity to empirical choices of hyperparameters such as model architecture, loss function, and optimisation algorithm.

Machine Translation Model Selection

Cannot find the paper you are looking for? You can Submit a new open access paper.