Search Results for author: Gunshi Gupta

Found 6 papers, 5 papers with code

La-MAML: Look-ahead Meta Learning for Continual Learning

3 code implementations ICML Workshop LifelongML 2020 Gunshi Gupta, Karmesh Yadav, Liam Paull

The continual learning problem involves training models with limited capacity to perform well on a set of an unknown number of sequentially arriving tasks.

Continual Learning Meta-Learning

Look-ahead Meta Learning for Continual Learning

2 code implementations NeurIPS 2020 Gunshi Gupta, Karmesh Yadav, Liam Paull

The continual learning problem involves training models with limited capacity to perform well on a set of an unknown number of sequentially arriving tasks.

Continual Learning Meta-Learning

ReLU to the Rescue: Improve Your On-Policy Actor-Critic with Positive Advantages

1 code implementation2 Jun 2023 Andrew Jesson, Chris Lu, Gunshi Gupta, Angelos Filos, Jakob Nicolaus Foerster, Yarin Gal

We show that the additive term is bounded proportional to the Lipschitz constant of the value function, which offers theoretical grounding for spectral normalization of critic weights.

Bayesian Inference Continuous Control +3

Unifying Variational Inference and PAC-Bayes for Supervised Learning that Scales

1 code implementation23 Oct 2019 Sanjay Thakur, Herke van Hoof, Gunshi Gupta, David Meger

PAC Bayes is a generalized framework which is more resistant to overfitting and that yields performance bounds that hold with arbitrarily high probability even on the unjustified extrapolations.

Variational Inference

Can Active Sampling Reduce Causal Confusion in Offline Reinforcement Learning?

1 code implementation28 Dec 2023 Gunshi Gupta, Tim G. J. Rudner, Rowan Thomas McAllister, Adrien Gaidon, Yarin Gal

To answer this question, we consider a set of tailored offline reinforcement learning datasets that exhibit causal ambiguity and assess the ability of active sampling techniques to reduce causal confusion at evaluation.

reinforcement-learning

Geometric Consistency for Self-Supervised End-to-End Visual Odometry

no code implementations11 Apr 2018 Ganesh Iyer, J. Krishna Murthy, Gunshi Gupta, K. Madhava Krishna, Liam Paull

We show that using a noisy teacher, which could be a standard VO pipeline, and by designing a loss term that enforces geometric consistency of the trajectory, we can train accurate deep models for VO that do not require ground-truth labels.

Visual Odometry

Cannot find the paper you are looking for? You can Submit a new open access paper.