Search Results for author: Jörg Bornschein

Found 5 papers, 2 papers with code

Towards Robust and Efficient Continual Language Learning

no code implementations11 Jul 2023 Adam Fisch, Amal Rannen-Triki, Razvan Pascanu, Jörg Bornschein, Angeliki Lazaridou, Elena Gribovskaya, Marc'Aurelio Ranzato

As the application space of language models continues to evolve, a natural question to ask is how we can quickly adapt models to new tasks.

Continual Learning

When Does Re-initialization Work?

no code implementations20 Jun 2022 Sheheryar Zaidi, Tudor Berariu, Hyunjik Kim, Jörg Bornschein, Claudia Clopath, Yee Whye Teh, Razvan Pascanu

However, when deployed alongside other carefully tuned regularization techniques, re-initialization methods offer little to no added benefit for generalization, although optimal generalization performance becomes less sensitive to the choice of learning rate and weight decay hyperparameters.

Data Augmentation Image Classification

ProSper -- A Python Library for Probabilistic Sparse Coding with Non-Standard Priors and Superpositions

no code implementations1 Aug 2019 Georgios Exarchakis, Jörg Bornschein, Abdul-Saboor Sheikh, Zhenwen Dai, Marc Henniges, Jakob Drefs, Jörg Lücke

The library widens the scope of dictionary learning approaches beyond implementations of standard approaches such as ICA, NMF or standard L1 sparse coding.

Dictionary Learning

Variational Memory Addressing in Generative Models

1 code implementation NeurIPS 2017 Jörg Bornschein, andriy mnih, Daniel Zoran, Danilo J. Rezende

Aiming to augment generative models with external memory, we interpret the output of a memory module with stochastic addressing as a conditional mixture distribution, where a read operation corresponds to sampling a discrete memory address and retrieving the corresponding content from memory.

Few-Shot Learning Representation Learning +1

Reweighted Wake-Sleep

2 code implementations11 Jun 2014 Jörg Bornschein, Yoshua Bengio

The wake-sleep algorithm relies on training not just the directed generative model but also a conditional generative model (the inference network) that runs backward from visible to latent, estimating the posterior distribution of latent given visible.

Cannot find the paper you are looking for? You can Submit a new open access paper.