Search Results for author: Timothée Lesort

Found 23 papers, 8 papers with code

Simple and Scalable Strategies to Continually Pre-train Large Language Models

1 code implementation13 Mar 2024 Adam Ibrahim, Benjamin Thérien, Kshitij Gupta, Mats L. Richter, Quentin Anthony, Timothée Lesort, Eugene Belilovsky, Irina Rish

In this work, we show that a simple and scalable combination of learning rate (LR) re-warming, LR re-decaying, and replay of previous data is sufficient to match the performance of fully re-training from scratch on all available data, as measured by the final loss and the average score on several language model (LM) evaluation benchmarks.

Continual Learning Language Modelling

Continual Learning Under Language Shift

no code implementations2 Nov 2023 Evangelia Gogoulou, Timothée Lesort, Magnus Boman, Joakim Nivre

The recent increase in data and model scale for language model pre-training has led to huge training costs.

Continual Learning Language Modelling

Continual Pre-Training of Large Language Models: How to (re)warm your model?

2 code implementations8 Aug 2023 Kshitij Gupta, Benjamin Thérien, Adam Ibrahim, Mats L. Richter, Quentin Anthony, Eugene Belilovsky, Irina Rish, Timothée Lesort

We study the warmup phase of models pre-trained on the Pile (upstream data, 300B tokens) as we continue to pre-train on SlimPajama (downstream data, 297B tokens), following a linear warmup and cosine decay schedule.

Language Modelling

Beyond Supervised Continual Learning: a Review

no code implementations30 Aug 2022 Benedikt Bagus, Alexander Gepperth, Timothée Lesort

Continual Learning (CL, sometimes also termed incremental learning) is a flavor of machine learning where the usual assumption of stationary data distribution is relaxed or omitted.

Continual Learning Incremental Learning +1

Challenging Common Assumptions about Catastrophic Forgetting

no code implementations10 Jul 2022 Timothée Lesort, Oleksiy Ostapenko, Diganta Misra, Md Rifat Arefin, Pau Rodríguez, Laurent Charlin, Irina Rish

In this paper, we study the progressive knowledge accumulation (KA) in DNNs trained with gradient-based algorithms in long sequences of tasks with data re-occurrence.

Continual Learning Memorization

Continual Feature Selection: Spurious Features in Continual Learning

no code implementations2 Mar 2022 Timothée Lesort

Our experiments highlight that continual learning algorithms face two related problems: (1) spurious features and (2) local spurious features.

Continual Learning feature selection

Continual Learning in Deep Networks: an Analysis of the Last Layer

no code implementations3 Jun 2021 Timothée Lesort, Thomas George, Irina Rish

Our analysis and results shed light on the dynamics of the output layer in continual learning scenarios and suggest a way of selecting the best type of output layer for a given scenario.

Continual Learning

Understanding Continual Learning Settings with Data Distribution Drift Analysis

no code implementations4 Apr 2021 Timothée Lesort, Massimo Caccia, Irina Rish

In this paper, we aim to identify and categorize different types of context drifts and potential assumptions about them, to better characterize various continual-learning scenarios.

BIG-bench Machine Learning Continual Learning

Continuum: Simple Management of Complex Continual Learning Scenarios

1 code implementation11 Feb 2021 Arthur Douillard, Timothée Lesort

Those drifts might cause interferences in the trained model and knowledge learned on previous states of the data distribution might be forgotten.

Continual Learning Management

Continual Learning: Tackling Catastrophic Forgetting in Deep Neural Networks with Replay Processes

no code implementations1 Jul 2020 Timothée Lesort

The replay processes make possible a compromise between optimizing the current learning objective and the past ones enabling learning without forgetting in sequences of tasks settings.

Continual Learning

Regularization Shortcomings for Continual Learning

no code implementations6 Dec 2019 Timothée Lesort, Andrei Stoian, David Filliat

In most machine learning algorithms, training data is assumed to be independent and identically distributed (iid).

Continual Learning Multi-Task Learning

DisCoRL: Continual Reinforcement Learning via Policy Distillation

no code implementations11 Jul 2019 René Traoré, Hugo Caselles-Dupré, Timothée Lesort, Te Sun, Guanghang Cai, Natalia Díaz-Rodríguez, David Filliat

In multi-task reinforcement learning there are two main challenges: at training time, the ability to learn different policies with a single model; at test time, inferring which of those policies applying without an external signal.

reinforcement-learning Reinforcement Learning (RL) +1

Continual Learning for Robotics: Definition, Framework, Learning Strategies, Opportunities and Challenges

no code implementations29 Jun 2019 Timothée Lesort, Vincenzo Lomonaco, Andrei Stoian, Davide Maltoni, David Filliat, Natalia Díaz-Rodríguez

An important challenge for machine learning is not necessarily finding solutions that work in the real world but rather finding stable algorithms that can learn in real world.

BIG-bench Machine Learning Continual Learning

Marginal Replay vs Conditional Replay for Continual Learning

no code implementations29 Oct 2018 Timothée Lesort, Alexander Gepperth, Andrei Stoian, David Filliat

We present a new replay-based method of continual classification learning that we term "conditional replay" which generates samples and labels together by sampling from a distribution conditioned on the class.

Classification Continual Learning +1

Training Discriminative Models to Evaluate Generative Ones

no code implementations28 Jun 2018 Timothée Lesort, Andrei Stoain, Jean-François Goudou, David Filliat

By comparing results with different generated datasets we are able to classify and compare generative models.

Generative Adversarial Network

State Representation Learning for Control: An Overview

1 code implementation12 Feb 2018 Timothée Lesort, Natalia Díaz-Rodríguez, Jean-François Goudou, David Filliat

State representation learning (SRL) focuses on a particular kind of representation learning where learned features are in low dimension, evolve through time, and are influenced by actions of an agent.

Representation Learning

Evaluation of generative networks through their data augmentation capacity

no code implementations ICLR 2018 Timothée Lesort, Florian Bordes, Jean-Francois Goudou, David Filliat

This mixture of real and generated data is thus used to train a classifier which is afterwards tested on a given labeled test dataset.

Data Augmentation

Unsupervised state representation learning with robotic priors: a robustness benchmark

no code implementations15 Sep 2017 Timothée Lesort, Mathieu Seurin, Xinrui Li, Natalia Díaz Rodríguez, David Filliat

We reproduce this simplification process using a neural network to build a low dimensional state representation of the world from images acquired by a robot.

Position Reinforcement Learning (RL) +2

Cannot find the paper you are looking for? You can Submit a new open access paper.