no code implementations • 1 May 2023 • Mélisande Teng, Amna Elmustafa, Benjamin Akera, Hugo Larochelle, David Rolnick
Climate change is a major driver of biodiversity loss, changing the geographic range and abundance of many species.
no code implementations • 15 Nov 2022 • Hattie Zhou, Azade Nova, Hugo Larochelle, Aaron Courville, Behnam Neyshabur, Hanie Sedghi
Large language models (LLMs) have shown increasing in-context learning capabilities through scaling up model and data size.
1 code implementation • 26 Jun 2022 • Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
With the success of large language models (LLMs) of code and their use as code assistants (e. g. Codex used in GitHub Copilot), techniques for introducing domain-specific knowledge in the prompt design process become important.
no code implementations • CVPR 2022 • Arman Afrasiyabi, Hugo Larochelle, Jean-François Lalonde, Christian Gagné
In image classification, it is common practice to train deep networks to extract a single feature vector per input image.
1 code implementation • 7 Mar 2022 • David Bieber, Rishab Goel, Daniel Zheng, Hugo Larochelle, Daniel Tarlow
This presents an interesting machine learning challenge: can we predict runtime errors in a "static" setting, where program execution is not possible?
1 code implementation • ICLR 2022 • Hattie Zhou, Ankit Vani, Hugo Larochelle, Aaron Courville
Forgetting is often seen as an unwanted characteristic in both human and machine learning.
1 code implementation • 10 Jan 2022 • Utku Evci, Vincent Dumoulin, Hugo Larochelle, Michael C. Mozer
We propose a method, Head-to-Toe probing (Head2Toe), that selects features from all layers of the source model to train a classification head for the target-domain.
no code implementations • 29 Sep 2021 • Utku Evci, Vincent Dumoulin, Hugo Larochelle, Michael Curtis Mozer
We propose a method, Head-to-Toe probing (Head2Toe), that selects features from all layers of the source model to train a classification head for the target-domain.
no code implementations • ICCV 2021 • Cristina Vasconcelos, Hugo Larochelle, Vincent Dumoulin, Rob Romijnders, Nicolas Le Roux, Ross Goroshin
We investigate the impact of aliasing on generalization in Deep Convolutional Networks and show that data augmentation schemes alone are unable to prevent it due to structural limitations in widely used architectures.
1 code implementation • NeurIPS 2021 • Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
Most learning-based approaches try to find a program that satisfies all examples at once.
1 code implementation • 14 May 2021 • Eleni Triantafillou, Hugo Larochelle, Richard Zemel, Vincent Dumoulin
Few-shot dataset generalization is a challenging variant of the well-studied few-shot classification problem where a diverse training set of several datasets is given, for the purpose of training an adaptable model that can then learn classes from new datasets using only a few examples.
1 code implementation • 6 Apr 2021 • Vincent Dumoulin, Neil Houlsby, Utku Evci, Xiaohua Zhai, Ross Goroshin, Sylvain Gelly, Hugo Larochelle
To bridge this gap, we perform a cross-family study of the best transfer and meta learners on both a large-scale meta-learning benchmark (Meta-Dataset, MD), and a transfer learning benchmark (Visual Task Adaptation Benchmark, VTAB).
no code implementations • 2 Mar 2021 • Prashanth Vijayaraghavan, Hugo Larochelle, Deb Roy
With growing role of social media in shaping public opinions and beliefs across the world, there has been an increased attention to identify and counter the problem of hate speech on social media.
no code implementations • 1 Feb 2021 • Cinjon Resnick, Or Litany, Cosmas Heiß, Hugo Larochelle, Joan Bruna, Kyunghyun Cho
We propose a self-supervised framework to learn scene representations from video that are automatically delineated into background, characters, and their animations.
no code implementations • 1 Jan 2021 • Nan Rosemary Ke, Olexa Bilaniuk, Anirudh Goyal, Stefan Bauer, Bernhard Schölkopf, Michael Curtis Mozer, Hugo Larochelle, Christopher Pal, Yoshua Bengio
Promising results have driven a recent surge of interest in continuous optimization methods for Bayesian network structure learning from observational data.
no code implementations • 1 Jan 2021 • Eleni Triantafillou, Vincent Dumoulin, Hugo Larochelle, Richard Zemel
We discover that fine-tuning on episodes of a particular shot can specialize the pre-trained model to solving episodes of that shot at the expense of performance on other shots, in agreement with a trade-off recently observed in the context of end-to-end episodic training.
no code implementations • 20 Nov 2020 • Cristina Vasconcelos, Hugo Larochelle, Vincent Dumoulin, Nicolas Le Roux, Ross Goroshin
Image pre-processing in the frequency domain has traditionally played a vital role in computer vision and was even part of the standard pipeline in the early days of deep learning.
no code implementations • 11 Nov 2020 • Cinjon Resnick, Or Litany, Hugo Larochelle, Joan Bruna, Kyunghyun Cho
We propose a self-supervised framework to learn scene representations from video that are automatically delineated into objects and background.
1 code implementation • NeurIPS 2020 • David Bieber, Charles Sutton, Hugo Larochelle, Daniel Tarlow
More practically, we evaluate these models on the task of learning to execute partial programs, as might arise if using the model as a heuristic function in program synthesis.
no code implementations • NeurIPS Workshop CAP 2020 • Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
The ability to adapt to unseen, local contexts is an important challenge that successful models of source code must overcome.
2 code implementations • ICML 2020 • William Fedus, Prajit Ramachandran, Rishabh Agarwal, Yoshua Bengio, Hugo Larochelle, Mark Rowland, Will Dabney
Experience replay is central to off-policy algorithms in deep reinforcement learning (RL), but there remain significant gaps in our understanding.
1 code implementation • NeurIPS 2020 • Daniel D. Johnson, Hugo Larochelle, Daniel Tarlow
In practice, edges are used both to represent intrinsic structure (e. g., abstract syntax trees of programs) and more abstract relations that aid reasoning for a downstream task (e. g., results of relevant program analyses).
no code implementations • 30 Jun 2020 • Samarth Sinha, Karsten Roth, Anirudh Goyal, Marzyeh Ghassemi, Hugo Larochelle, Animesh Garg
Deep Neural Networks have shown great promise on a variety of downstream applications; but their ability to adapt and generalize to new data and tasks remains a challenge.
1 code implementation • ICLR 2021 • Lu Liu, William Hamilton, Guodong Long, Jing Jiang, Hugo Larochelle
We consider the problem of multi-domain few-shot image classification, where unseen classes and examples come from diverse data sources.
Ranked #1 on
Few-Shot Image Classification
on Meta-Dataset Rank
no code implementations • 27 Mar 2020 • Joelle Pineau, Philippe Vincent-Lamarre, Koustuv Sinha, Vincent Larivière, Alina Beygelzimer, Florence d'Alché-Buc, Emily Fox, Hugo Larochelle
Reproducibility, that is obtaining similar results as presented in a paper or talk, using the same code and data (when available), is a necessary step to verify the reliability of research findings.
no code implementations • 26 Mar 2020 • Disha Shrivastava, Hugo Larochelle, Daniel Tarlow
The ability to adapt to unseen, local contexts is an important challenge that successful models of source code must overcome.
3 code implementations • NeurIPS 2020 • Tong Che, Ruixiang Zhang, Jascha Sohl-Dickstein, Hugo Larochelle, Liam Paull, Yuan Cao, Yoshua Bengio
To make that practical, we show that sampling from this modified density can be achieved by sampling in latent space according to an energy-based model induced by the sum of the latent prior log-density and the discriminator output score.
1 code implementation • 10 Mar 2020 • Samarth Sinha, Homanga Bharadhwaj, Anirudh Goyal, Hugo Larochelle, Animesh Garg, Florian Shkurti
Although deep learning models have achieved state-of-the-art performance on a number of vision tasks, generalization over high dimensional multi-modal data, and reliable predictive uncertainty estimation are still active areas of research.
2 code implementations • NeurIPS 2020 • Samarth Sinha, Animesh Garg, Hugo Larochelle
We propose to augment the train-ing of CNNs by controlling the amount of high frequency information propagated within the CNNs as training progresses, by convolving the output of a CNN feature map of each layer with a Gaussian kernel.
1 code implementation • 28 Feb 2020 • William Fedus, Dibya Ghosh, John D. Martin, Marc G. Bellemare, Yoshua Bengio, Hugo Larochelle
Our study provides a clear empirical link between catastrophic interference and sample efficiency in reinforcement learning.
no code implementations • 28 Nov 2019 • Vishal Jain, William Fedus, Hugo Larochelle, Doina Precup, Marc G. Bellemare
Empirically, we find that these techniques improve the performance of a baseline deep reinforcement learning agent applied to text-based games.
no code implementations • ICML 2020 • Samarth Sinha, Han Zhang, Anirudh Goyal, Yoshua Bengio, Hugo Larochelle, Augustus Odena
Recent work by Brock et al. (2018) suggests that Generative Adversarial Networks (GANs) benefit disproportionately from large mini-batch sizes.
2 code implementations • 2 Oct 2019 • Nan Rosemary Ke, Olexa Bilaniuk, Anirudh Goyal, Stefan Bauer, Hugo Larochelle, Bernhard Schölkopf, Michael C. Mozer, Chris Pal, Yoshua Bengio
Promising results have driven a recent surge of interest in continuous optimization methods for Bayesian network structure learning from observational data.
no code implementations • 25 Sep 2019 • Gabriel Huang, Hugo Larochelle, Simon Lacoste-Julien
We argue that the widely used Omniglot and miniImageNet benchmarks are too simple because their class semantics do not vary across episodes, which defeats their intended purpose of evaluating few-shot classification methods.
no code implementations • ICLR 2019 • Anirudh Goyal, Riashat Islam, DJ Strouse, Zafarali Ahmed, Hugo Larochelle, Matthew Botvinick, Yoshua Bengio, Sergey Levine
In new environments, this model can then identify novel subgoals for further exploration, guiding the agent through a sequence of potential decision states and through new regions of the state space.
no code implementations • ICLR Workshop DeepGenStruct 2019 • Laurent Dinh, Jascha Sohl-Dickstein, Hugo Larochelle, Razvan Pascanu
Flow based models such as Real NVP are an extremely powerful approach to density estimation.
12 code implementations • ICLR 2020 • Eleni Triantafillou, Tyler Zhu, Vincent Dumoulin, Pascal Lamblin, Utku Evci, Kelvin Xu, Ross Goroshin, Carles Gelada, Kevin Swersky, Pierre-Antoine Manzagol, Hugo Larochelle
Few-shot classification refers to learning a classifier for new classes given only a few examples.
Ranked #7 on
Few-Shot Image Classification
on Meta-Dataset Rank
1 code implementation • 22 Feb 2019 • Gabriel Huang, Hugo Larochelle, Simon Lacoste-Julien
We show that several popular few-shot learning benchmarks can be solved with varying degrees of success without using support set Labels at Test-time (LT).
1 code implementation • ICLR 2020 • William Fedus, Carles Gelada, Yoshua Bengio, Marc G. Bellemare, Hugo Larochelle
Reinforcement learning (RL) typically defines a discount factor as part of the Markov Decision Process.
2 code implementations • 1 Feb 2019 • Nolan Bard, Jakob N. Foerster, Sarath Chandar, Neil Burch, Marc Lanctot, H. Francis Song, Emilio Parisotto, Vincent Dumoulin, Subhodeep Moitra, Edward Hughes, Iain Dunning, Shibl Mourad, Hugo Larochelle, Marc G. Bellemare, Michael Bowling
From the early days of computing, games have been important testbeds for studying how well machines can do sophisticated decision making.
no code implementations • 30 Jan 2019 • Anirudh Goyal, Riashat Islam, Daniel Strouse, Zafarali Ahmed, Matthew Botvinick, Hugo Larochelle, Yoshua Bengio, Sergey Levine
In new environments, this model can then identify novel subgoals for further exploration, guiding the agent through a sequence of potential decision states and through new regions of the state space.
1 code implementation • 12 Nov 2018 • Ankesh Anand, Eugene Belilovsky, Kyle Kastner, Hugo Larochelle, Aaron Courville
We explore blindfold (question-only) baselines for Embodied Question Answering.
1 code implementation • ICLR 2020 • Massimo Caccia, Lucas Caccia, William Fedus, Hugo Larochelle, Joelle Pineau, Laurent Charlin
Generating high-quality text with sufficient diversity is essential for a wide range of Natural Language Generation (NLG) tasks.
no code implementations • ICLR 2019 • Anirudh Goyal, Philemon Brakel, William Fedus, Soumye Singhal, Timothy Lillicrap, Sergey Levine, Hugo Larochelle, Yoshua Bengio
In many environments only a tiny subset of all states yield high reward.
8 code implementations • ICLR 2018 • Mengye Ren, Eleni Triantafillou, Sachin Ravi, Jake Snell, Kevin Swersky, Joshua B. Tenenbaum, Hugo Larochelle, Richard S. Zemel
To address this paradigm, we propose novel extensions of Prototypical Networks (Snell et al., 2017) that are augmented with the ability to use unlabeled examples when producing prototypes.
no code implementations • 26 Feb 2018 • Valentin Thomas, Emmanuel Bengio, William Fedus, Jules Pondard, Philippe Beaudoin, Hugo Larochelle, Joelle Pineau, Doina Precup, Yoshua Bengio
It has been postulated that a good representation is one that disentangles the underlying explanatory factors of variation.
no code implementations • NeurIPS 2017 • Manasi Vartak, Arvind Thiagarajan, Conrado Miranda, Jeshua Bratman, Hugo Larochelle
Matrix factorization (MF) is one of the most popular techniques for product recommendation, but is known to suffer from serious cold-start problems.
no code implementations • 29 Nov 2017 • Simon Brodeur, Ethan Perez, Ankesh Anand, Florian Golemo, Luca Celotti, Florian Strub, Jean Rouat, Hugo Larochelle, Aaron Courville
We introduce HoME: a Household Multimodal Environment for artificial agents to learn from vision, audio, semantics, physics, and interaction with objects and other agents, all within a realistic context.
no code implementations • 3 Jul 2017 • Bart van Merriënboer, Amartya Sanyal, Hugo Larochelle, Yoshua Bengio
We propose a generalization of neural network sequence models.
3 code implementations • NeurIPS 2017 • Harm de Vries, Florian Strub, Jérémie Mary, Hugo Larochelle, Olivier Pietquin, Aaron Courville
It is commonly assumed that language refers to high-level visual concepts while leaving low-level visual processing unaffected.
4 code implementations • CVPR 2017 • Harm de Vries, Florian Strub, Sarath Chandar, Olivier Pietquin, Hugo Larochelle, Aaron Courville
Our key contribution is the collection of a large-scale dataset consisting of 150K human-played games with a total of 800K visual question-answer pairs on 66K images.
no code implementations • 18 Jul 2016 • Mohammad Havaei, Nicolas Guizard, Hugo Larochelle, Pierre-Marc Jodoin
In this chapter, we provide a survey of CNN methods applied to medical imaging with a focus on brain pathology segmentation.
no code implementations • 24 May 2016 • Sarath Chandar, Sungjin Ahn, Hugo Larochelle, Pascal Vincent, Gerald Tesauro, Yoshua Bengio
In this paper, we explore a form of hierarchical memory network, which can be considered as a hybrid between hard and soft attention memory networks.
no code implementations • 12 May 2016 • Anna Rohrbach, Atousa Torabi, Marcus Rohrbach, Niket Tandon, Christopher Pal, Hugo Larochelle, Aaron Courville, Bernt Schiele
In addition we also collected and aligned movie scripts used in prior work and compare the two sources of descriptions.
3 code implementations • 7 May 2016 • Benigno Uria, Marc-Alexandre Côté, Karol Gregor, Iain Murray, Hugo Larochelle
We present Neural Autoregressive Distribution Estimation (NADE) models, which are neural network architectures applied to the problem of unsupervised distribution and density estimation.
no code implementations • 27 Mar 2016 • Loris Bazzani, Hugo Larochelle, Lorenzo Torresani
In this work, we propose a spatiotemporal attentional model that learns where to look in a video directly from human fixation data.
no code implementations • 18 Mar 2016 • Stanislas Lauly, Yin Zheng, Alexandre Allauzen, Hugo Larochelle
We present an approach based on feed-forward neural networks for learning the distribution of textual documents.
28 code implementations • 31 Dec 2015 • Anders Boesen Lindbo Larsen, Søren Kaae Sønderby, Hugo Larochelle, Ole Winther
We present an autoencoder that leverages learned representations to better measure similarities in data space.
2 code implementations • 24 Nov 2015 • Amjad Almahairi, Nicolas Ballas, Tim Cooijmans, Yin Zheng, Hugo Larochelle, Aaron Courville
The low-capacity sub-networks are applied across most of the input, but also provide a guide to select a few portions of the input on which to apply the high-capacity sub-networks.
no code implementations • 5 Oct 2015 • Mohammad Havaei, Hugo Larochelle, Philippe Poulin, Pierre-Marc Jodoin
Purpose: In this paper, we investigate a framework for interactive brain tumor segmentation which, at its core, treats the problem of interactive brain tumor segmentation as a machine learning problem.
no code implementations • 21 Jul 2015 • Alex Auvolat, Sarath Chandar, Pascal Vincent, Hugo Larochelle, Yoshua Bengio
Efficient Maximum Inner Product Search (MIPS) is an important task that has a wide applicability in recommendation systems and classification with a large number of classes.
35 code implementations • 28 May 2015 • Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, Victor Lempitsky
Our approach is directly inspired by the theory on domain adaptation suggesting that, for effective domain transfer to be achieved, predictions must be made based on features that cannot discriminate between the training (source) and test (target) domains.
Ranked #2 on
Domain Adaptation
on Synth Digits-to-SVHN
14 code implementations • 13 May 2015 • Mohammad Havaei, Axel Davy, David Warde-Farley, Antoine Biard, Aaron Courville, Yoshua Bengio, Chris Pal, Pierre-Marc Jodoin, Hugo Larochelle
Finally, we explore a cascade architecture in which the output of a basic CNN is treated as an additional source of information for a subsequent CNN.
Ranked #1 on
Brain Tumor Segmentation
on BRATS-2013 leaderboard
2 code implementations • 27 Apr 2015 • Sarath Chandar, Mitesh M. Khapra, Hugo Larochelle, Balaraman Ravindran
CCA based approaches learn a joint representation by maximizing correlation of the views when projected to the common subspace.
1 code implementation • 3 Mar 2015 • Atousa Torabi, Christopher Pal, Hugo Larochelle, Aaron Courville
DVS is an audio narration describing the visual elements and actions in a movie for the visually impaired.
5 code implementations • ICCV 2015 • Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
In this context, we propose an approach that successfully takes into account both the local and global temporal structure of videos to produce descriptions.
16 code implementations • 12 Feb 2015 • Mathieu Germain, Karol Gregor, Iain Murray, Hugo Larochelle
There has been a lot of recent interest in designing neural network models to estimate a distribution from a set of examples.
Ranked #4 on
Density Estimation
on UCI GAS
1 code implementation • 9 Feb 2015 • Marc-Alexandre Côté, Hugo Larochelle
We present a mathematical construction for the restricted Boltzmann machine (RBM) that doesn't require specifying the number of hidden units.
1 code implementation • 15 Dec 2014 • Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand
We propose a training objective that implements this idea in the context of a neural network, whose hidden layer is trained to be predictive of the classification task, but uninformative as to the domain of the input.
no code implementations • 13 Sep 2014 • Yin Zheng, Yu-Jin Zhang, Hugo Larochelle
Second, we propose a deep extension of our model and provide an efficient way of training the deep model.
no code implementations • CVPR 2014 • Yin Zheng, Yu-Jin Zhang, Hugo Larochelle
Topic modeling based on latent Dirichlet allocation (LDA) has been a framework of choice to deal with multimodal data, such as in image annotation tasks.
no code implementations • NeurIPS 2014 • Sarath Chandar A P, Stanislas Lauly, Hugo Larochelle, Mitesh M. Khapra, Balaraman Ravindran, Vikas Raykar, Amrita Saha
Cross-language learning allows us to use training data from one language to build models for a different language.
no code implementations • 4 Feb 2014 • Alexandre Lacoste, Hugo Larochelle, François Laviolette, Mario Marchand
One of the most tedious tasks in the application of machine learning is model selection, i. e. hyperparameter selection.
no code implementations • 8 Jan 2014 • Stanislas Lauly, Alex Boulanger, Hugo Larochelle
Recent work on learning multilingual word representations usually relies on the use of word-level alignements (e. g. infered with the help of GIZA++) between translated sentences, in order to align the word embeddings in different languages.
no code implementations • 7 Oct 2013 • Benigno Uria, Iain Murray, Hugo Larochelle
We can thus use the most convenient model for each inference task at hand, and ensembles of such models with different orderings are immediately available.
Ranked #9 on
Image Generation
on Binarized MNIST
no code implementations • NeurIPS 2013 • Benigno Uria, Iain Murray, Hugo Larochelle
We introduce RNADE, a new model for joint density estimation of real-valued vectors.
no code implementations • 23 May 2013 • Yin Zheng, Yu-Jin Zhang, Hugo Larochelle
Topic modeling based on latent Dirichlet allocation (LDA) has been a framework of choice to perform scene recognition and annotation.
no code implementations • NeurIPS 2012 • Hugo Larochelle, Stanislas Lauly
We describe a new model for learning meaningful representations of text documents from an unlabeled collection of documents.
4 code implementations • NeurIPS 2012 • Jasper Snoek, Hugo Larochelle, Ryan P. Adams
In this work, we consider the automatic tuning problem within the framework of Bayesian optimization, in which a learning algorithm's generalization performance is modeled as a sample from a Gaussian process (GP).
Ranked #188 on
Image Classification
on CIFAR-10
no code implementations • NeurIPS 2010 • Hugo Larochelle, Geoffrey E. Hinton
We describe a model based on a Boltzmann machine with third-order connections that can learn how to accumulate information about a shape over several fixations.