At the same time, advances in approximate Bayesian methods have made posterior approximation for flexible neural network models practical.
SOTA for Multi-Armed Bandits on Mushroom
We show that this form of adversarial training converges to a degenerate global minimum, wherein small curvature artifacts near the data points obfuscate a linear approximation of the loss.
We show that generating English Wikipedia articles can be approached as a multi- document summarization of source documents.
In this work, we study how depthwise separable convolutions can be applied to neural machine translation.
#21 best model for Machine Translation on WMT2014 English-German
We describe a new training methodology for generative adversarial networks.
#2 best model for Image Generation on CIFAR-10