no code implementations • CVPR 2024 • Xiaoqi Wang, Wenbin He, Xiwei Xuan, Clint Sebastian, Jorge Piazentin Ono, Xin Li, Sima Behpour, Thang Doan, Liang Gou, Han Wei Shen, Liu Ren
The main challenge in open-vocabulary image segmentation now lies in accurately classifying these segments into text-defined categories.
2 code implementations • 10 Mar 2024 • Thang Doan, Sima Behpour, Xin Li, Wenbin He, Liang Gou, Liu Ren
Few-shot Class-Incremental Learning (FSCIL) poses the challenge of retaining prior knowledge while learning from limited new data streams, all without overfitting.
class-incremental learning
Few-Shot Class-Incremental Learning
+1
2 code implementations • 25 Jun 2023 • Thang Doan, Xin Li, Sima Behpour, Wenbin He, Liang Gou, Liu Ren
We argue that this contextual information should already be embedded within the known classes.
1 code implementation • 18 Nov 2022 • Jean-Baptiste Gaya, Thang Doan, Lucas Caccia, Laure Soulier, Ludovic Denoyer, Roberta Raileanu
We introduce Continual Subspace of Policies (CSP), a new approach that incrementally builds a subspace of policies for training a reinforcement learning agent on a sequence of tasks.
no code implementations • 20 Feb 2022 • Thang Doan, Seyed Iman Mirzadeh, Mehrdad Farajtabar
A growing body of research in continual learning focuses on the catastrophic forgetting problem.
no code implementations • 14 Feb 2021 • Bonnie Li, Vincent François-Lavet, Thang Doan, Joelle Pineau
We consider the problem of generalization in reinforcement learning where visual aspects of the observations might differ, e. g. when there are different backgrounds or change in contrast, brightness, etc.
no code implementations • ICLR 2021 • Wonseok Jeon, Chen-Yang Su, Paul Barde, Thang Doan, Derek Nowrouzezahrai, Joelle Pineau
Inverse Reinforcement Learning (IRL) aims to facilitate a learner's ability to imitate expert behavior by acquiring reward functions that explain the expert's decisions.
2 code implementations • 7 Oct 2020 • Thang Doan, Mehdi Bennani, Bogdan Mazoure, Guillaume Rabusseau, Pierre Alquier
Continual learning (CL) is a setting in which an agent has to learn from an incoming stream of data during its entire lifetime.
1 code implementation • 21 Jun 2020 • Mehdi Abbana Bennani, Thang Doan, Masashi Sugiyama
In this framework, we prove that OGD is robust to Catastrophic Forgetting then derive the first generalisation bound for SGD and OGD for Continual Learning.
1 code implementation • NeurIPS 2020 • Bogdan Mazoure, Remi Tachet des Combes, Thang Doan, Philip Bachman, R. Devon Hjelm
We begin with the hypothesis that a model-free agent whose representations are predictive of properties of future states (beyond expected rewards) will be more capable of solving and adapting to new RL problems.
no code implementations • 7 Feb 2020 • Bogdan Mazoure, Thang Doan, Tianyu Li, Vladimir Makarenkov, Joelle Pineau, Doina Precup, Guillaume Rabusseau
We propose a general framework for policy representation for reinforcement learning tasks.
no code implementations • 17 Sep 2019 • Thang Doan, Bogdan Mazoure, Moloud Abdar, Audrey Durand, Joelle Pineau, R. Devon Hjelm
Continuous control tasks in reinforcement learning are important because they provide an important framework for learning in high-dimensional state spaces with deceptive rewards, where the agent can easily become trapped into suboptimal solutions.
no code implementations • 5 Jul 2019 • Srinivas Venkattaramanujam, Eric Crawford, Thang Doan, Doina Precup
Goal-conditioned policies are used in order to break down complex reinforcement learning (RL) problems by using subgoals, which can be defined either in state space or in a latent feature space.
1 code implementation • 16 May 2019 • Bogdan Mazoure, Thang Doan, Audrey Durand, R. Devon Hjelm, Joelle Pineau
The ability to discover approximately optimal policies in domains with sparse rewards is crucial to applying reinforcement learning (RL) in many real-world scenarios.
1 code implementation • ICLR 2019 • Isabela Albuquerque, João Monteiro, Thang Doan, Breandan Considine, Tiago Falk, Ioannis Mitliagkas
Recent literature has demonstrated promising results for training Generative Adversarial Networks by employing a set of discriminators, in contrast to the traditional game involving one generator against a single adversary.
no code implementations • 17 Jan 2019 • Thang Doan, Neil Veira, Saibal Ray, Brian Keng
The GAN is thus used in tandem with the RNN module in a pipeline alternating between basket generation and customer state updating steps.
no code implementations • WS 2018 • Bogdan Mazoure, Thang Doan, Saibal Ray
Generative models have recently experienced a surge in popularity due to the development of more efficient training algorithms and increasing computational power.
3 code implementations • 31 Jul 2018 • Thang Doan, Joao Monteiro, Isabela Albuquerque, Bogdan Mazoure, Audrey Durand, Joelle Pineau, R. Devon Hjelm
We argue that less expressive discriminators are smoother and have a general coarse grained view of the modes map, which enforces the generator to cover a wide portion of the data distribution support.
1 code implementation • 13 May 2018 • Thang Doan, Bogdan Mazoure, Clare Lyle
Distributional reinforcement learning (distributional RL) has seen empirical success in complex Markov Decision Processes (MDPs) in the setting of nonlinear function approximation.
1 code implementation • 6 Dec 2017 • Peter Henderson, Thang Doan, Riashat Islam, David Meger
Policy gradient methods have had great success in solving continuous control tasks, yet the stochastic nature of such problems makes deterministic value estimation difficult.