ICLR 2018

Deep Bayesian Bandits Showdown: An Empirical Comparison of Bayesian Deep Networks for Thompson Sampling

ICLR 2018 tensorflow/models

At the same time, advances in approximate Bayesian methods have made posterior approximation for flexible neural network models practical.

DECISION MAKING MULTI-ARMED BANDITS

Scalable Private Learning with PATE

ICLR 2018 tensorflow/models

To address those concerns, one promising approach is Private Aggregation of Teacher Ensembles, or PATE, which transfers to a "student" model the knowledge of an ensemble of "teacher" models, with intuitive privacy provided by training teachers on disjoint data and strong privacy guaranteed by noisy aggregation of teachers' answers.

Generating Wikipedia by Summarizing Long Sequences

ICLR 2018 tensorflow/tensor2tensor

We show that generating English Wikipedia articles can be approached as a multi- document summarization of source documents.

DOCUMENT SUMMARIZATION MULTI-DOCUMENT SUMMARIZATION

Discrete Autoencoders for Sequence Models

ICLR 2018 tensorflow/tensor2tensor

We propose to improve the representation in sequence models by augmenting current approaches with an autoencoder that is forced to compress the sequence through an intermediate discrete latent space.

LANGUAGE MODELLING MACHINE TRANSLATION

Unsupervised Machine Translation Using Monolingual Corpora Only

ICLR 2018 facebookresearch/MUSE

By learning to reconstruct in both languages from this shared feature space, the model effectively learns to translate without using any labeled data.

UNSUPERVISED MACHINE TRANSLATION

Word Translation Without Parallel Data

ICLR 2018 facebookresearch/MUSE

We finally describe experiments on the English-Esperanto low-resource language pair, on which there only exists a limited amount of parallel data, to show the potential impact of our method in fully unsupervised machine translation.

UNSUPERVISED MACHINE TRANSLATION WORD EMBEDDINGS

Model compression via distillation and quantization

ICLR 2018 NervanaSystems/distiller

Deep neural networks (DNNs) continue to make significant advances, solving tasks from image classification to translation or reinforcement learning.

MODEL COMPRESSION

Mixed Precision Training

ICLR 2018 NVIDIA/DeepRecommender

Using this approach, we can reduce the memory consumption of deep learning models by nearly 2x.