At the same time, advances in approximate Bayesian methods have made posterior approximation for flexible neural network models practical.
SOTA for Multi-Armed Bandits on Mushroom
We show that this form of adversarial training converges to a degenerate global minimum, wherein small curvature artifacts near the data points obfuscate a linear approximation of the loss.
In this work, we study how depthwise separable convolutions can be applied to neural machine translation.
#22 best model for Machine Translation on WMT2014 English-German
We show that generating English Wikipedia articles can be approached as a multi- document summarization of source documents.
Recurrent neural networks (RNNs), such as long short-term memory networks (LSTMs), serve as a fundamental building block for many sequence learning tasks, including machine translation, language modeling, and question answering.
#8 best model for Language Modelling on WikiText-2
We describe a new training methodology for generative adversarial networks.
#2 best model for Image Generation on CIFAR-10