no code implementations • 9 Jan 2025 • Maximilian Alber, Stephan Tietz, Jonas Dippel, Timo Milbich, Timothée Lesort, Panos Korfiatis, Moritz Krügener, Beatriz Perez Cancer, Neelay Shah, Alexander Möllers, Philipp Seegerer, Alexandra Carpen-Amarie, Kai Standvoss, Gabriel Dernbach, Edwin de Jong, Simon Schallenberg, Andreas Kunft, Helmut Hoffer von Ankershoffen, Gavin Schaeferle, Patrick Duffy, Matt Redlon, Philipp Jurmeister, David Horst, Lukas Ruff, Klaus-Robert Müller, Frederick Klauschen, Andrew Norgan
Recent advances in digital pathology have demonstrated the effectiveness of foundation models across diverse applications.
1 code implementation • 13 Mar 2024 • Adam Ibrahim, Benjamin Thérien, Kshitij Gupta, Mats L. Richter, Quentin Anthony, Timothée Lesort, Eugene Belilovsky, Irina Rish
In this work, we show that a simple and scalable combination of learning rate (LR) re-warming, LR re-decaying, and replay of previous data is sufficient to match the performance of fully re-training from scratch on all available data, as measured by the final loss and the average score on several language model (LM) evaluation benchmarks.
no code implementations • 2 Nov 2023 • Evangelia Gogoulou, Timothée Lesort, Magnus Boman, Joakim Nivre
The recent increase in data and model scale for language model pre-training has led to huge training costs.
2 code implementations • 8 Aug 2023 • Kshitij Gupta, Benjamin Thérien, Adam Ibrahim, Mats L. Richter, Quentin Anthony, Eugene Belilovsky, Irina Rish, Timothée Lesort
We study the warmup phase of models pre-trained on the Pile (upstream data, 300B tokens) as we continue to pre-train on SlimPajama (downstream data, 297B tokens), following a linear warmup and cosine decay schedule.
no code implementations • 30 Aug 2022 • Benedikt Bagus, Alexander Gepperth, Timothée Lesort
Continual Learning (CL, sometimes also termed incremental learning) is a flavor of machine learning where the usual assumption of stationary data distribution is relaxed or omitted.
no code implementations • 10 Jul 2022 • Timothée Lesort, Oleksiy Ostapenko, Diganta Misra, Md Rifat Arefin, Pau Rodríguez, Laurent Charlin, Irina Rish
In this paper, we study the progressive knowledge accumulation (KA) in DNNs trained with gradient-based algorithms in long sequences of tasks with data re-occurrence.
no code implementations • 2 Mar 2022 • Timothée Lesort
Our experiments highlight that continual learning algorithms face two related problems: (1) spurious features and (2) local spurious features.
3 code implementations • 2 Aug 2021 • Fabrice Normandin, Florian Golemo, Oleksiy Ostapenko, Pau Rodriguez, Matthew D Riemer, Julio Hurtado, Khimya Khetarpal, Ryan Lindeborg, Lucas Cecchi, Timothée Lesort, Laurent Charlin, Irina Rish, Massimo Caccia
We propose a taxonomy of settings, where each setting is described as a set of assumptions.
no code implementations • 3 Jun 2021 • Timothée Lesort, Thomas George, Irina Rish
Our analysis and results shed light on the dynamics of the output layer in continual learning scenarios and suggest a way of selecting the best type of output layer for a given scenario.
no code implementations • 4 Apr 2021 • Timothée Lesort, Massimo Caccia, Irina Rish
In this paper, we aim to identify and categorize different types of context drifts and potential assumptions about them, to better characterize various continual-learning scenarios.
1 code implementation • 11 Feb 2021 • Arthur Douillard, Timothée Lesort
Those drifts might cause interferences in the trained model and knowledge learned on previous states of the data distribution might be forgotten.
no code implementations • 1 Jul 2020 • Timothée Lesort
The replay processes make possible a compromise between optimizing the current learning objective and the past ones enabling learning without forgetting in sequences of tasks settings.
no code implementations • 6 Dec 2019 • Timothée Lesort, Andrei Stoian, David Filliat
In most machine learning algorithms, training data is assumed to be independent and identically distributed (iid).
no code implementations • 11 Jul 2019 • René Traoré, Hugo Caselles-Dupré, Timothée Lesort, Te Sun, Guanghang Cai, Natalia Díaz-Rodríguez, David Filliat
In multi-task reinforcement learning there are two main challenges: at training time, the ability to learn different policies with a single model; at test time, inferring which of those policies applying without an external signal.
no code implementations • 29 Jun 2019 • Timothée Lesort, Vincenzo Lomonaco, Andrei Stoian, Davide Maltoni, David Filliat, Natalia Díaz-Rodríguez
An important challenge for machine learning is not necessarily finding solutions that work in the real world but rather finding stable algorithms that can learn in real world.
no code implementations • 11 Jun 2019 • René Traoré, Hugo Caselles-Dupré, Timothée Lesort, Te Sun, Natalia Díaz-Rodríguez, David Filliat
We focus on the problem of teaching a robot to solve tasks presented sequentially, i. e., in a continual learning scenario.
5 code implementations • 24 Jan 2019 • Antonin Raffin, Ashley Hill, René Traoré, Timothée Lesort, Natalia Díaz-Rodríguez, David Filliat
Scaling end-to-end reinforcement learning to control real robots from vision presents a series of challenges, in particular in terms of sample efficiency.
1 code implementation • ICLR 2019 • Timothée Lesort, Hugo Caselles-Dupré, Michael Garcia-Ortiz, Andrei Stoian, David Filliat
We experiment with sequential tasks on three commonly used benchmarks for Continual Learning (MNIST, Fashion MNIST and CIFAR10).
no code implementations • 29 Oct 2018 • Timothée Lesort, Alexander Gepperth, Andrei Stoian, David Filliat
We present a new replay-based method of continual classification learning that we term "conditional replay" which generates samples and labels together by sampling from a distribution conditioned on the class.
5 code implementations • 25 Sep 2018 • Antonin Raffin, Ashley Hill, René Traoré, Timothée Lesort, Natalia Díaz-Rodríguez, David Filliat
State representation learning aims at learning compact representations from raw observations in robotics and control applications.
no code implementations • 28 Jun 2018 • Timothée Lesort, Andrei Stoain, Jean-François Goudou, David Filliat
By comparing results with different generated datasets we are able to classify and compare generative models.
1 code implementation • 12 Feb 2018 • Timothée Lesort, Natalia Díaz-Rodríguez, Jean-François Goudou, David Filliat
State representation learning (SRL) focuses on a particular kind of representation learning where learned features are in low dimension, evolve through time, and are influenced by actions of an agent.
no code implementations • ICLR 2018 • Timothée Lesort, Florian Bordes, Jean-Francois Goudou, David Filliat
This mixture of real and generated data is thus used to train a classifier which is afterwards tested on a given labeled test dataset.
no code implementations • 15 Sep 2017 • Timothée Lesort, Mathieu Seurin, Xinrui Li, Natalia Díaz Rodríguez, David Filliat
We reproduce this simplification process using a neural network to build a low dimensional state representation of the world from images acquired by a robot.