Search Results for author: Massimo Caccia

Found 21 papers, 15 papers with code

Towards Compute-Optimal Transfer Learning

no code implementations25 Apr 2023 Massimo Caccia, Alexandre Galashov, Arthur Douillard, Amal Rannen-Triki, Dushyant Rao, Michela Paganini, Laurent Charlin, Marc'Aurelio Ranzato, Razvan Pascanu

The field of transfer learning is undergoing a significant shift with the introduction of large pretrained models which have demonstrated strong adaptability to a variety of downstream tasks.

Computational Efficiency Continual Learning +1

Task-Agnostic Continual Reinforcement Learning: Gaining Insights and Overcoming Challenges

2 code implementations28 May 2022 Massimo Caccia, Jonas Mueller, Taesup Kim, Laurent Charlin, Rasool Fakoor

We pose two hypotheses: (1) task-agnostic methods might provide advantages in settings with limited data, computation, or high dimensionality, and (2) faster adaptation may be particularly beneficial in continual learning settings, helping to mitigate the effects of catastrophic forgetting.

Continual Learning Continuous Control +3

Continual Learning via Local Module Composition

1 code implementation NeurIPS 2021 Oleksiy Ostapenko, Pau Rodriguez, Massimo Caccia, Laurent Charlin

We introduce local module composition (LMC), an approach to modular CL where each module is provided a local structural component that estimates a module's relevance to the input.

Continual Learning Transfer Learning

Pretrained Language Model in Continual Learning: A Comparative Study

no code implementations ICLR 2022 Tongtong Wu, Massimo Caccia, Zhuang Li, Yuan-Fang Li, Guilin Qi, Gholamreza Haffari

In this paper, we thoroughly compare the continual learning performance over the combination of 5 PLMs and 4 veins of CL methods on 3 benchmarks in 2 typical incremental settings.

Continual Learning Language Modelling

Understanding Continual Learning Settings with Data Distribution Drift Analysis

no code implementations4 Apr 2021 Timothée Lesort, Massimo Caccia, Irina Rish

In this paper, we aim to identify and categorize different types of context drifts and potential assumptions about them, to better characterize various continual-learning scenarios.

BIG-bench Machine Learning Continual Learning

Beyond Trivial Counterfactual Explanations with Diverse Valuable Explanations

2 code implementations ICCV 2021 Pau Rodriguez, Massimo Caccia, Alexandre Lacoste, Lee Zamparo, Issam Laradji, Laurent Charlin, David Vazquez

Explainability for machine learning models has gained considerable attention within the research community given the importance of deploying more reliable machine-learning systems.

Attribute BIG-bench Machine Learning +2

Beyond Trivial Counterfactual Generations with Diverse Valuable Explanations

no code implementations1 Jan 2021 Pau Rodriguez, Massimo Caccia, Alexandre Lacoste, Lee Zamparo, Issam H. Laradji, Laurent Charlin, David Vazquez

In computer vision applications, most methods explain models by displaying the regions in the input image that they focus on for their prediction, but it is difficult to improve models based on these explanations since they do not indicate why the model fail.

Attribute counterfactual +1

CVPR 2020 Continual Learning in Computer Vision Competition: Approaches, Results, Current Challenges and Future Directions

1 code implementation14 Sep 2020 Vincenzo Lomonaco, Lorenzo Pellegrini, Pau Rodriguez, Massimo Caccia, Qi She, Yu Chen, Quentin Jodelet, Ruiping Wang, Zheda Mai, David Vazquez, German I. Parisi, Nikhil Churamani, Marc Pickett, Issam Laradji, Davide Maltoni

In the last few years, we have witnessed a renewed and fast-growing interest in continual learning with deep neural networks with the shared objective of making current AI systems more adaptive, efficient and autonomous.

Benchmarking Continual Learning

Online Continual Learning with Maximal Interfered Retrieval

2 code implementations NeurIPS 2019 Rahaf Aljundi, Eugene Belilovsky, Tinne Tuytelaars, Laurent Charlin, Massimo Caccia, Min Lin, Lucas Page-Caccia

Methods based on replay, either generative or from a stored memory, have been shown to be effective approaches for continual learning, matching or exceeding the state of the art in a number of standard benchmarks.

Class Incremental Learning Retrieval

Online Learned Continual Compression with Adaptive Quantization Modules

1 code implementation ICML 2020 Lucas Caccia, Eugene Belilovsky, Massimo Caccia, Joelle Pineau

We show how to use discrete auto-encoders to effectively address this challenge and introduce Adaptive Quantization Modules (AQM) to control variation in the compression ability of the module at any given stage of learning.

Continual Learning Quantization

Online Learned Continual Compression with Stacked Quantization Modules

no code implementations25 Sep 2019 Lucas Caccia, Eugene Belilovsky, Massimo Caccia, Joelle Pineau

We first replace the episodic memory used in Experience Replay with SQM, leading to significant gains on standard continual learning benchmarks using a fixed memory budget.

Continual Learning Quantization

Online Continual Learning with Maximally Interfered Retrieval

1 code implementation11 Aug 2019 Rahaf Aljundi, Lucas Caccia, Eugene Belilovsky, Massimo Caccia, Min Lin, Laurent Charlin, Tinne Tuytelaars

Methods based on replay, either generative or from a stored memory, have been shown to be effective approaches for continual learning, matching or exceeding the state of the art in a number of standard benchmarks.

Continual Learning Retrieval

Language GANs Falling Short

1 code implementation ICLR 2020 Massimo Caccia, Lucas Caccia, William Fedus, Hugo Larochelle, Joelle Pineau, Laurent Charlin

Generating high-quality text with sufficient diversity is essential for a wide range of Natural Language Generation (NLG) tasks.

Text Generation

Option Pricing and Hedging for Discrete Time Autoregressive Hidden Markov Model

1 code implementation7 Jul 2017 Massimo Caccia, Bruno Rémillard

In this paper we solve the discrete time mean-variance hedging problem when asset returns follow a multivariate autoregressive hidden Markov model.

Time Series Time Series Analysis

Cannot find the paper you are looking for? You can Submit a new open access paper.