1 code implementation • ICML 2020 • Sai Krishna Gottipati, Boris Sattarov, Sufeng. Niu, Hao-Ran Wei, Yashaswi Pathak, Shengchao Liu, Simon Blackburn, Karam Thomas, Connor Coley, Jian Tang, Sarath Chandar, Yoshua Bengio
In this work, we propose a novel reinforcement learning (RL) setup for drug discovery that addresses this challenge by embedding the concept of synthetic accessibility directly into the de novo compound design system.
1 code implementation • SIGDIAL (ACL) 2021 • Prasanna Parthasarathi, Joelle Pineau, Sarath Chandar
Predicting the next utterance in dialogue is contingent on encoding of users’ input text to generate appropriate and relevant response in data-driven approaches.
no code implementations • 20 Aug 2023 • Hadi Nekoei, Xutong Zhao, Janarthanan Rajendran, Miao Liu, Sarath Chandar
In this work, we show empirically that state-of-the-art ZSC algorithms have poor performance when paired with agents trained with different learning methods, and they require millions of interaction samples to adapt to these new partners.
no code implementations • 31 Jul 2023 • Gonçalo Mordido, Pranshu Malviya, Aristide Baratin, Sarath Chandar
The Lookahead optimizer improves the training stability of deep neural networks by having a set of fast weights that "look ahead" to guide the descent direction.
1 code implementation • 18 Jul 2023 • Pranshu Malviya, Gonçalo Mordido, Aristide Baratin, Reza Babanezhad Harikandeh, Jerry Huang, Simon Lacoste-Julien, Razvan Pascanu, Sarath Chandar
Adaptive gradient-based optimizers, particularly Adam, have left their mark in training large-scale deep learning models.
no code implementations • 30 Jun 2023 • Jarrid Rector-Brooks, Kanika Madan, Moksh Jain, Maksym Korablyov, Cheng-Hao Liu, Sarath Chandar, Nikolay Malkin, Yoshua Bengio
Generative flow networks (GFlowNets) are amortized variational inference algorithms that treat sampling from a distribution over compositional objects as a sequential decision-making problem with a learnable action policy.
no code implementations • 24 May 2023 • Amirhossein Kazemnejad, Mehdi Rezagholizadeh, Prasanna Parthasarathi, Sarath Chandar
We propose a systematic framework to measure parametric knowledge utilization in PLMs.
no code implementations • 22 May 2023 • Abdelrahman Zayed, Goncalo Mordido, Samira Shabanian, Sarath Chandar
In this work, we investigate the role of attention, a widely-used technique in current state-of-the-art NLP models, in the propagation of social biases.
1 code implementation • 16 Mar 2023 • Xutong Zhao, Yangchen Pan, Chenjun Xiao, Sarath Chandar, Janarthanan Rajendran
Efficient exploration is critical in cooperative deep Multi-Agent Reinforcement Learning (MARL).
no code implementations • 15 Mar 2023 • Ali Rahimi-Kalahroudi, Janarthanan Rajendran, Ida Momennejad, Harm van Seijen, Sarath Chandar
This is challenging for deep-learning-based world models due to catastrophic forgetting.
Model-based Reinforcement Learning
reinforcement-learning
+1
no code implementations • 6 Feb 2023 • Hadi Nekoei, Akilesh Badrinaaraayanan, Amit Sinha, Mohammad Amini, Janarthanan Rajendran, Aditya Mahajan, Sarath Chandar
In our proposed method, when one agent updates its policy, other agents are allowed to update their policies as well, but at a slower rate.
no code implementations • 20 Nov 2022 • Abdelrahman Zayed, Prasanna Parthasarathi, Goncalo Mordido, Hamid Palangi, Samira Shabanian, Sarath Chandar
The fairness achieved by our method surpasses that of data augmentation on three text classification datasets, using no more than half of the examples in the augmented dataset.
no code implementations • 18 Nov 2022 • Gonçalo Mordido, Sébastien Henwood, Sarath Chandar, François Leduc-Primeau
In this work, we show that applying sharpness-aware training, by optimizing for both the loss value and loss sharpness, significantly improves robustness to noisy hardware at inference time without relying on any assumptions about the target hardware.
no code implementations • 11 Nov 2022 • Gabriele Prato, Yale Song, Janarthanan Rajendran, R Devon Hjelm, Neel Joshi, Sarath Chandar
We show that our method is successful at enabling vision transformers to encode the temporal component of video data.
no code implementations • 9 Nov 2022 • Louis Clouâtre, Prasanna Parthasarathi, Amal Zouaq, Sarath Chandar
In this work, we replicate a study on the importance of local structure, and the relative unimportance of global structure, in a multilingual setting.
no code implementations • 9 Nov 2022 • Louis Clouâtre, Prasanna Parthasarathi, Amal Zouaq, Sarath Chandar
However, this transfer is not universal, with many languages not currently understood by multilingual approaches.
1 code implementation • 27 Oct 2022 • Enamundram Naga Karthik, Anne Kerbrat, Pierre Labauge, Tobias Granberg, Jason Talbott, Daniel S. Reich, Massimo Filippi, Rohit Bakshi, Virginie Callot, Sarath Chandar, Julien Cohen-Adad
Segmentation of Multiple Sclerosis (MS) lesions is a challenging problem.
1 code implementation • 3 Aug 2022 • Simon Guiroy, Christopher Pal, Gonçalo Mordido, Sarath Chandar
Specifically, we analyze the evolution, during meta-training, of the neural activations at each hidden layer, on a small set of unlabelled support examples from a single task of the target tasks distribution, as this constitutes a minimal and justifiably accessible information from the target problem.
no code implementations • 10 Jul 2022 • Shagun Sodhani, Mojtaba Faramarzi, Sanket Vaibhav Mehta, Pranshu Malviya, Mohamed Abdelsalam, Janarthanan Janarthanan, Sarath Chandar
Following these different classes of learning algorithms, we discuss the commonly used evaluation benchmarks and metrics for lifelong learning (Chapter 6) and wrap up with a discussion of future challenges and important research directions in Chapter 7.
1 code implementation • 25 Apr 2022 • Yi Wan, Ali Rahimi-Kalahroudi, Janarthanan Rajendran, Ida Momennejad, Sarath Chandar, Harm van Seijen
We empirically validate these insights in the case of linear function approximation by demonstrating that a modified version of linear Dyna achieves effective adaptation to local changes.
Model-based Reinforcement Learning
reinforcement-learning
+1
no code implementations • 1 Feb 2022 • Amir Ardalan Kalantari, Mohammad Amini, Sarath Chandar, Doina Precup
Much of recent Deep Reinforcement Learning success is owed to the neural architecture's potential to learn and use effective internal representations of the world.
1 code implementation • 16 Dec 2021 • Sanket Vaibhav Mehta, Darshan Patil, Sarath Chandar, Emma Strubell
The lifelong learning paradigm in machine learning is an attractive alternative to the more prominent isolated learning scheme not only due to its resemblance to biological learning but also its potential to reduce energy waste by obviating excessive model re-training.
no code implementations • 13 Oct 2021 • Gabriele Prato, Simon Guiroy, Ethan Caballero, Irina Rish, Sarath Chandar
Empirical science of neural scaling laws is a rapidly growing area of significant importance to the future of machine learning, particularly in the light of recent breakthroughs achieved by large-scale pre-trained models such as GPT-3, CLIP and DALL-e.
no code implementations • 29 Sep 2021 • Simon Guiroy, Christopher Pal, Sarath Chandar
To this end, we empirically show that as meta-training progresses, a model's generalization to a target distribution of novel tasks can be estimated by analysing the dynamics of its neural activations.
no code implementations • 10 Aug 2021 • Andreas Madsen, Siva Reddy, Sarath Chandar
Neural networks for NLP are becoming increasingly complex and widespread, and there is a growing concern if these models are responsible to use.
no code implementations • Findings (ACL) 2022 • Louis Clouatre, Prasanna Parthasarathi, Amal Zouaq, Sarath Chandar
Recent research analyzing the sensitivity of natural language understanding models to word-order perturbations has shown that neural models are surprisingly insensitive to the order of words.
1 code implementation • 20 Jun 2021 • Prasanna Parthasarathi, Joelle Pineau, Sarath Chandar
Predicting the next utterance in dialogue is contingent on encoding of users' input text to generate appropriate and relevant response in data-driven approaches.
1 code implementation • SIGDIAL (ACL) 2021 • Prasanna Parthasarathi, Mohamed Abdelsalam, Joelle Pineau, Sarath Chandar
Neural models trained for next utterance generation in dialogue task learn to mimic the n-gram sequences in the training set with training objectives like negative log-likelihood (NLL) or cross-entropy.
2 code implementations • ICLR 2022 • Paul-Aymeric McRae, Prasanna Parthasarathi, Mahmoud Assran, Sarath Chandar
Popular approaches for minimizing loss in data-driven learning often involve an abstraction or an explicit retention of the history of gradients for efficient parameter updates.
1 code implementation • 11 May 2021 • Pranshu Malviya, Balaraman Ravindran, Sarath Chandar
We also show that our method performs better than several state-of-the-art methods in lifelong learning on complex datasets with a large number of tasks.
1 code implementation • Findings (ACL) 2021 • Steven Y. Feng, Varun Gangal, Jason Wei, Sarath Chandar, Soroush Vosoughi, Teruko Mitamura, Eduard Hovy
In this paper, we present a comprehensive and unifying survey of data augmentation for NLP by summarizing the literature in a structured manner.
2 code implementations • 4 Mar 2021 • Hadi Nekoei, Akilesh Badrinaaraayanan, Aaron Courville, Sarath Chandar
Its large strategy space makes it a desirable environment for lifelong RL tasks.
no code implementations • 1 Jan 2021 • Gabriele Prato, Sarath Chandar
This includes left out classes from the same dataset, as well as entire datasets never trained on.
1 code implementation • CVPR 2021 • Mohamed Abdelsalam, Mojtaba Faramarzi, Shagun Sodhani, Sarath Chandar
We develop a standardized benchmark that enables evaluating models on the IIRC setup.
1 code implementation • 8 Oct 2020 • Sai Krishna Gottipati, Yashaswi Pathak, Rohan Nuttall, Sahir, Raviteja Chunduru, Ahmed Touati, Sriram Ganapathi Subramanian, Matthew E. Taylor, Sarath Chandar
Reinforcement learning (RL) algorithms typically deal with maximizing the expected cumulative return (discounted or undiscounted, finite or infinite horizon).
no code implementations • Findings (ACL) 2021 • Louis Clouatre, Philippe Trempe, Amal Zouaq, Sarath Chandar
They however scale with man-hours and high-quality data.
Ranked #10 on
Link Prediction
on WN18RR
(using extra training data)
1 code implementation • 24 Aug 2020 • Prasanna Parthasarathi, Joelle Pineau, Sarath Chandar
To bridge this gap in evaluation, we propose designing a set of probing tasks to evaluate dialogue models.
no code implementations • 18 Jul 2020 • Evan Racah, Sarath Chandar
Unsupervised extraction of objects from low-level visual data is an important goal for further progress in machine learning.
2 code implementations • NeurIPS 2020 • Harm van Seijen, Hadi Nekoei, Evan Racah, Sarath Chandar
For example, the common single-task sample-efficiency metric conflates improvements due to model-based learning with various other aspects, such as representation learning, making it difficult to assess true progress on model-based RL.
Model-based Reinforcement Learning
Reinforcement Learning (RL)
+1
no code implementations • ICML Workshop LifelongML 2020 • Touraj Laleh, Mojtaba Faramarzi, Irina Rish, Sarath Chandar
Most proposed approaches for this issue try to compensate for the effects of parameter updates in the batch incremental setup in which the training model visits a lot of samples for several epochs.
1 code implementation • 14 Jun 2020 • Mojtaba Faramarzi, Mohammad Amini, Akilesh Badrinaaraayanan, Vikas Verma, Sarath Chandar
Our approach improves the robustness of CNN models against the manifold intrusion problem that may occur in other state-of-the-art mixing approaches.
1 code implementation • 26 Apr 2020 • Sai Krishna Gottipati, Boris Sattarov, Sufeng. Niu, Yashaswi Pathak, Hao-Ran Wei, Shengchao Liu, Karam M. J. Thomas, Simon Blackburn, Connor W. Coley, Jian Tang, Sarath Chandar, Yoshua Bengio
Over the last decade, there has been significant progress in the field of machine learning for de novo drug design, particularly in deep generative models.
1 code implementation • ACL 2019 • Gabriele Prato, Mathieu Duchesneau, Sarath Chandar, Alain Tapp
A lot of work has been done in the field of image compression via machine learning, but not much attention has been given to the compression of natural language.
1 code implementation • ACL 2019 • Chinnadhurai Sankar, Sandeep Subramanian, Christopher Pal, Sarath Chandar, Yoshua Bengio
Neural generative models have been become increasingly popular when building conversational agents.
no code implementations • WS 2019 • Vardaan Pahuja, Jie Fu, Sarath Chandar, Christopher J. Pal
In current formulations of such networks only the parameters of the neural modules and/or the order of their execution is learned.
2 code implementations • 1 Feb 2019 • Nolan Bard, Jakob N. Foerster, Sarath Chandar, Neil Burch, Marc Lanctot, H. Francis Song, Emilio Parisotto, Vincent Dumoulin, Subhodeep Moitra, Edward Hughes, Iain Dunning, Shibl Mourad, Hugo Larochelle, Marc G. Bellemare, Michael Bowling
From the early days of computing, games have been important testbeds for studying how well machines can do sophisticated decision making.
2 code implementations • 22 Jan 2019 • Sarath Chandar, Chinnadhurai Sankar, Eugene Vorontsov, Samira Ebrahimi Kahou, Yoshua Bengio
Modelling long-term dependencies is a challenge for recurrent neural networks.
2 code implementations • 26 Nov 2018 • Khimya Khetarpal, Shagun Sodhani, Sarath Chandar, Doina Precup
To achieve general artificial intelligence, reinforcement learning (RL) agents should learn not only to optimize returns for one specific task but also to constantly build more complex skills and scaffold their knowledge about the world, without forgetting what has already been learned.
no code implementations • 16 Nov 2018 • Shagun Sodhani, Sarath Chandar, Yoshua Bengio
Both these models are proposed in the context of feedforward networks and we evaluate the feasibility of using them for recurrent networks.
no code implementations • 17 May 2018 • Ghulam Ahmed Ansari, Sagar J P, Sarath Chandar, Balaraman Ravindran
Text-based games are suitable test-beds for designing agents that can learn by interaction with the environment in the form of natural language text.
1 code implementation • 31 Jan 2018 • Amrita Saha, Vardaan Pahuja, Mitesh M. Khapra, Karthik Sankaranarayanan, Sarath Chandar
Further, unlike existing large scale QA datasets which contain simple questions that can be answered from a single tuple, the questions in our dialogs require a larger subgraph of the KG.
no code implementations • 20 Jan 2018 • Iulian V. Serban, Chinnadhurai Sankar, Mathieu Germain, Saizheng Zhang, Zhouhan Lin, Sandeep Subramanian, Taesup Kim, Michael Pieper, Sarath Chandar, Nan Rosemary Ke, Sai Rajeswar, Alexandre de Brebisson, Jose M. R. Sotelo, Dendi Suhubdy, Vincent Michalski, Alexandre Nguyen, Joelle Pineau, Yoshua Bengio
We present MILABOT: a deep reinforcement learning chatbot developed by the Montreal Institute for Learning Algorithms (MILA) for the Amazon Alexa Prize competition.
no code implementations • 7 Sep 2017 • Iulian V. Serban, Chinnadhurai Sankar, Mathieu Germain, Saizheng Zhang, Zhouhan Lin, Sandeep Subramanian, Taesup Kim, Michael Pieper, Sarath Chandar, Nan Rosemary Ke, Sai Rajeshwar, Alexandre de Brebisson, Jose M. R. Sotelo, Dendi Suhubdy, Vincent Michalski, Alexandre Nguyen, Joelle Pineau, Yoshua Bengio
By applying reinforcement learning to crowdsourced data and real-world user interactions, the system has been trained to select an appropriate response from the models in its ensemble.
no code implementations • 30 Jan 2017 • Caglar Gulcehre, Sarath Chandar, Yoshua Bengio
We use discrete addressing for read/write operations which helps to substantially to reduce the vanishing gradient problem with very long sequences.
4 code implementations • CVPR 2017 • Harm de Vries, Florian Strub, Sarath Chandar, Olivier Pietquin, Hugo Larochelle, Aaron Courville
Our key contribution is the collection of a large-scale dataset consisting of 150K human-played games with a total of 800K visual question-answer pairs on 66K images.
no code implementations • 30 Jun 2016 • Caglar Gulcehre, Sarath Chandar, Kyunghyun Cho, Yoshua Bengio
We investigate the mechanisms and effects of learning to read and write into a memory through experiments on Facebook bAbI tasks using both a feedforward and GRUcontroller.
Ranked #5 on
Question Answering
on bAbi
no code implementations • COLING 2016 • Amrita Saha, Mitesh M. Khapra, Sarath Chandar, Janarthanan Rajendran, Kyunghyun Cho
However, there is no parallel training data available between X and Y but, training data is available between X & Z and Z & Y (as is often the case in many real world applications).
no code implementations • 24 May 2016 • Sarath Chandar, Sungjin Ahn, Hugo Larochelle, Pascal Vincent, Gerald Tesauro, Yoshua Bengio
In this paper, we explore a form of hierarchical memory network, which can be considered as a hybrid between hard and soft attention memory networks.
1 code implementation • ACL 2016 • Iulian Vlad Serban, Alberto García-Durán, Caglar Gulcehre, Sungjin Ahn, Sarath Chandar, Aaron Courville, Yoshua Bengio
Over the past decade, large-scale supervised learning corpora have enabled machine learning researchers to make substantial advances.
1 code implementation • NAACL 2016 • Janarthanan Rajendran, Mitesh M. Khapra, Sarath Chandar, Balaraman Ravindran
In this work, we address a real-world scenario where no direct parallel data is available between two views of interest (say, $V_1$ and $V_2$) but parallel data is available between each of these views and a pivot view ($V_3$).
no code implementations • 10 Oct 2015 • P. Prasanna, Sarath Chandar, Balaraman Ravindran
In this paper, we propose TSEB, a Thompson Sampling based algorithm with adaptive exploration bonus that aims to solve the problem with tighter PAC guarantees, while being cautious on the regret as well.
no code implementations • 28 Jul 2015 • Sridhar Mahadevan, Sarath Chandar
In this paper, we introduce a new approach to capture analogies in continuous word representations, based on modeling not just individual word vectors, but rather the subspaces spanned by groups of words.
no code implementations • 21 Jul 2015 • Alex Auvolat, Sarath Chandar, Pascal Vincent, Hugo Larochelle, Yoshua Bengio
Efficient Maximum Inner Product Search (MIPS) is an important task that has a wide applicability in recommendation systems and classification with a large number of classes.
2 code implementations • 27 Apr 2015 • Sarath Chandar, Mitesh M. Khapra, Hugo Larochelle, Balaraman Ravindran
CCA based approaches learn a joint representation by maximizing correlation of the views when projected to the common subspace.