Search Results for author: Riashat Islam

Found 26 papers, 8 papers with code

PcLast: Discovering Plannable Continuous Latent States

no code implementations6 Nov 2023 Anurag Koul, Shivakanth Sujit, Shaoru Chen, Ben Evans, Lili Wu, Byron Xu, Rajan Chari, Riashat Islam, Raihan Seraj, Yonathan Efroni, Lekan Molu, Miro Dudik, John Langford, Alex Lamb

Goal-conditioned planning benefits from learned low-dimensional representations of rich, high-dimensional observations.

Ignorance is Bliss: Robust Control via Information Gating

no code implementations NeurIPS 2023 Manan Tomar, Riashat Islam, Matthew E. Taylor, Sergey Levine, Philip Bachman

We propose \textit{information gating} as a way to learn parsimonious representations that identify the minimal information required for a task.

Inductive Bias Q-Learning

Offline Policy Optimization in RL with Variance Regularizaton

no code implementations29 Dec 2022 Riashat Islam, Samarth Sinha, Homanga Bharadhwaj, Samin Yeasar Arnob, Zhuoran Yang, Animesh Garg, Zhaoran Wang, Lihong Li, Doina Precup

Learning policies from fixed offline datasets is a key challenge to scale up reinforcement learning (RL) algorithms towards practical applications.

Continuous Control Offline RL +1

Behavior Prior Representation learning for Offline Reinforcement Learning

1 code implementation2 Nov 2022 Hongyu Zang, Xin Li, Jie Yu, Chen Liu, Riashat Islam, Remi Tachet des Combes, Romain Laroche

Our method, Behavior Prior Representation (BPR), learns state representations with an easy-to-integrate objective based on behavior cloning of the dataset: we first learn a state representation by mimicking actions from the dataset, and then train a policy on top of the fixed representation, using any off-the-shelf Offline RL algorithm.

Offline RL reinforcement-learning +2

Discrete Factorial Representations as an Abstraction for Goal Conditioned Reinforcement Learning

no code implementations1 Nov 2022 Riashat Islam, Hongyu Zang, Anirudh Goyal, Alex Lamb, Kenji Kawaguchi, Xin Li, Romain Laroche, Yoshua Bengio, Remi Tachet des Combes

Goal-conditioned reinforcement learning (RL) is a promising direction for training agents that are capable of solving multiple tasks and reach a diverse set of objectives.

reinforcement-learning Reinforcement Learning (RL)

Agent-Controller Representations: Principled Offline RL with Rich Exogenous Information

1 code implementation31 Oct 2022 Riashat Islam, Manan Tomar, Alex Lamb, Yonathan Efroni, Hongyu Zang, Aniket Didolkar, Dipendra Misra, Xin Li, Harm van Seijen, Remi Tachet des Combes, John Langford

We find that contemporary representation learning techniques can fail on datasets where the noise is a complex and time dependent process, which is prevalent in practical applications.

Offline RL Reinforcement Learning (RL) +1

Guaranteed Discovery of Control-Endogenous Latent States with Multi-Step Inverse Models

no code implementations17 Jul 2022 Alex Lamb, Riashat Islam, Yonathan Efroni, Aniket Didolkar, Dipendra Misra, Dylan Foster, Lekan Molu, Rajan Chari, Akshay Krishnamurthy, John Langford

In many sequential decision-making tasks, the agent is not able to model the full complexity of the world, which consists of multitudes of relevant and irrelevant information.

Decision Making

Importance of Empirical Sample Complexity Analysis for Offline Reinforcement Learning

no code implementations31 Dec 2021 Samin Yeasar Arnob, Riashat Islam, Doina Precup

We hypothesize that empirically studying the sample complexity of offline reinforcement learning (RL) is crucial for the practical applications of RL in the real world.

Offline RL reinforcement-learning +1

Offline Policy Optimization with Variance Regularization

no code implementations1 Jan 2021 Riashat Islam, Samarth Sinha, Homanga Bharadhwaj, Samin Yeasar Arnob, Zhuoran Yang, Zhaoran Wang, Animesh Garg, Lihong Li, Doina Precup

Learning policies from fixed offline datasets is a key challenge to scale up reinforcement learning (RL) algorithms towards practical applications.

Continuous Control Offline RL +1

Marginalized State Distribution Entropy Regularization in Policy Optimization

no code implementations11 Dec 2019 Riashat Islam, Zafarali Ahmed, Doina Precup

Entropy regularization is used to get improved optimization performance in reinforcement learning tasks.

Continuous Control

Doubly Robust Off-Policy Actor-Critic Algorithms for Reinforcement Learning

no code implementations11 Dec 2019 Riashat Islam, Raihan Seraj, Samin Yeasar Arnob, Doina Precup

Furthermore, in cases where the reward function is stochastic that can lead to high variance, doubly robust critic estimation can improve performance under corrupted, stochastic reward signals, indicating its usefulness for robust and safe reinforcement learning.

Continuous Control reinforcement-learning +2

Entropy Regularization with Discounted Future State Distribution in Policy Gradient Methods

no code implementations11 Dec 2019 Riashat Islam, Raihan Seraj, Pierre-Luc Bacon, Doina Precup

In this work, we propose exploration in policy gradient methods based on maximizing entropy of the discounted future state distribution.

Policy Gradient Methods

Off-Policy Policy Gradient Algorithms by Constraining the State Distribution Shift

no code implementations16 Nov 2019 Riashat Islam, Komal K. Teru, Deepak Sharma, Joelle Pineau

This data distribution shift between current and past samples can significantly impact the performance of most modern off-policy based policy optimization algorithms.

Continuous Control Reinforcement Learning (RL)

Transfer Learning by Modeling a Distribution over Policies

no code implementations9 Jun 2019 Disha Shrivastava, Eeshan Gunesh Dhekane, Riashat Islam

Exploration and adaptation to new tasks in a transfer learning setup is a central challenge in reinforcement learning.

reinforcement-learning Reinforcement Learning (RL) +1

Transfer and Exploration via the Information Bottleneck

no code implementations ICLR 2019 Anirudh Goyal, Riashat Islam, DJ Strouse, Zafarali Ahmed, Hugo Larochelle, Matthew Botvinick, Yoshua Bengio, Sergey Levine

In new environments, this model can then identify novel subgoals for further exploration, guiding the agent through a sequence of potential decision states and through new regions of the state space.

InfoBot: Transfer and Exploration via the Information Bottleneck

no code implementations30 Jan 2019 Anirudh Goyal, Riashat Islam, Daniel Strouse, Zafarali Ahmed, Matthew Botvinick, Hugo Larochelle, Yoshua Bengio, Sergey Levine

In new environments, this model can then identify novel subgoals for further exploration, guiding the agent through a sequence of potential decision states and through new regions of the state space.

Exploring Restart Distributions

no code implementations27 Nov 2018 Arash Tavakoli, Vitaly Levdik, Riashat Islam, Christopher M. Smith, Petar Kormushev

We consider the generic approach of using an experience memory to help exploration by adapting a restart distribution.

Bayesian Policy Gradients via Alpha Divergence Dropout Inference

1 code implementation6 Dec 2017 Peter Henderson, Thang Doan, Riashat Islam, David Meger

Policy gradient methods have had great success in solving continuous control tasks, yet the stochastic nature of such problems makes deterministic value estimation difficult.

Continuous Control Policy Gradient Methods

Alpha-Divergences in Variational Dropout

no code implementations12 Nov 2017 Bogdan Mazoure, Riashat Islam

We investigate the use of alternative divergences to Kullback-Leibler (KL) in variational inference(VI), based on the Variational Dropout \cite{kingma2015}.

Variational Inference

Deep Reinforcement Learning that Matters

4 code implementations19 Sep 2017 Peter Henderson, Riashat Islam, Philip Bachman, Joelle Pineau, Doina Precup, David Meger

In recent years, significant progress has been made in solving challenging problems across various domains using deep reinforcement learning (RL).

Atari Games Continuous Control +2

Reproducibility of Benchmarked Deep Reinforcement Learning Tasks for Continuous Control

1 code implementation10 Aug 2017 Riashat Islam, Peter Henderson, Maziar Gomrokchi, Doina Precup

We investigate and discuss: the significance of hyper-parameters in policy gradients for continuous control, general variance in the algorithms, and reproducibility of reported results.

Continuous Control Policy Gradient Methods +2

Deep Bayesian Active Learning with Image Data

5 code implementations ICML 2017 Yarin Gal, Riashat Islam, Zoubin Ghahramani

In this paper we combine recent advances in Bayesian deep learning into the active learning framework in a practical way.

Active Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.