Search Results for author: François Fleuret

Found 49 papers, 21 papers with code

σ-GPTs: A New Approach to Autoregressive Models

no code implementations15 Apr 2024 Arnaud Pannatier, Evann Courdier, François Fleuret

Autoregressive models, such as the GPT family, use a fixed order, usually left-to-right, to generate sequences.

Language Modelling

Multi-Lattice Sampling of Quantum Field Theories via Neural Operator-based Flows

no code implementations1 Jan 2024 Bálint Máté, François Fleuret

In particular, we propose to approximate a time-dependent operator $\mathcal V_t$ whose time integral provides a mapping between the functional distributions of the free theory $[\mathcal D\phi(x)] \mathcal Z_0^{-1} e^{-\mathcal S_{0}[\phi(x)]}$ and of the target theory $[\mathcal D\phi(x)]\mathcal Z^{-1}e^{-\mathcal S[\phi(x)]}$.

Operator learning

PAUMER: Patch Pausing Transformer for Semantic Segmentation

no code implementations1 Nov 2023 Evann Courdier, Prabhu Teja Sivaprasad, François Fleuret

We study the problem of improving the efficiency of segmentation transformers by using disparate amounts of computation for different parts of the image.

Segmentation Semantic Segmentation

Faster Causal Attention Over Large Sequences Through Sparse Flash Attention

1 code implementation1 Jun 2023 Matteo Pagliardini, Daniele Paliotta, Martin Jaggi, François Fleuret

While many works have proposed schemes to sparsify the attention patterns and reduce the computational overhead of self-attention, those are often limited by implementations concerns and end up imposing a simple and static structure over the attention matrix.

16k 8k +1

Graph Neural Networks Go Forward-Forward

no code implementations10 Feb 2023 Daniele Paliotta, Mathieu Alain, Bálint Máté, François Fleuret

We present the Graph Forward-Forward (GFF) algorithm, an extension of the Forward-Forward procedure to graphs, able to handle features distributed over a graph's nodes.

Graph Property Prediction Property Prediction

Learning Interpolations between Boltzmann Densities

1 code implementation18 Jan 2023 Bálint Máté, François Fleuret

We introduce a training objective for continuous normalizing flows that can be used in the absence of samples but in the presence of an energy function.

Deformations of Boltzmann Distributions

no code implementations25 Oct 2022 Bálint Máté, François Fleuret

Consider a one-parameter family of Boltzmann distributions $p_t(x) = \tfrac{1}{Z_t}e^{-S_t(x)}$.

Inference from Real-World Sparse Measurements

no code implementations20 Oct 2022 Arnaud Pannatier, Kyle Matoba, François Fleuret

Notably, our model reduces the root mean square error (RMSE) for wind nowcasting from 9. 24 to 7. 98 and for heat diffusion tasks from 0. 126 to 0. 084.

Weather Forecasting

Pareto Manifold Learning: Tackling multiple tasks via ensembles of single-task models

1 code implementation18 Oct 2022 Nikolaos Dimitriadis, Pascal Frossard, François Fleuret

In Multi-Task Learning (MTL), tasks may compete and limit the performance achieved on each other, rather than guiding the optimization to a solution, superior to all its single-task trained counterparts.

Image Classification Multi-Task Learning +1

Transformers are Sample-Efficient World Models

1 code implementation1 Sep 2022 Vincent Micheli, Eloi Alonso, François Fleuret

Deep reinforcement learning agents are notoriously sample inefficient, which considerably limits their application to real-world problems.

Atari Games 100k reinforcement-learning +1

Flowification: Everything is a Normalizing Flow

1 code implementation30 May 2022 Bálint Máté, Samuel Klein, Tobias Golling, François Fleuret

On the other hand, neural networks only perform a forward pass on the input, there is neither a notion of an inverse of a neural network nor is there one of its likelihood contribution.

Density Estimation

The Theoretical Expressiveness of Maxpooling

no code implementations2 Mar 2022 Kyle Matoba, Nikolaos Dimitriadis, François Fleuret

Over the decade since deep neural networks became state of the art image classifiers there has been a tendency towards less use of max pooling: the function that takes the largest of nearby pixels in an image.

MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned

no code implementations17 Feb 2022 Anssi Kanervisto, Stephanie Milani, Karolis Ramanauskas, Nicholay Topin, Zichuan Lin, Junyou Li, Jianing Shi, Deheng Ye, Qiang Fu, Wei Yang, Weijun Hong, Zhongyue Huang, Haicheng Chen, Guangjun Zeng, Yue Lin, Vincent Micheli, Eloi Alonso, François Fleuret, Alexander Nikulin, Yury Belousov, Oleg Svidchenko, Aleksei Shpilman

With this in mind, we hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.

Borrowing from yourself: Faster future video segmentation with partial channel update

1 code implementation11 Feb 2022 Evann Courdier, François Fleuret

Semantic segmentation is a well-addressed topic in the computer vision literature, but the design of fast and accurate video processing networks remains challenging.

Future prediction Semantic Segmentation +2

Agree to Disagree: Diversity through Disagreement for Better Transferability

1 code implementation9 Feb 2022 Matteo Pagliardini, Martin Jaggi, François Fleuret, Sai Praneeth Karimireddy

This behavior can hinder the transferability of trained models by (i) favoring the learning of simpler but spurious features -- present in the training data but absent from the test data -- and (ii) by only leveraging a small subset of predictive features.

Out of Distribution (OOD) Detection

Efficient Wind Speed Nowcasting with GPU-Accelerated Nearest Neighbors Algorithm

1 code implementation20 Dec 2021 Arnaud Pannatier, Ricardo Picatoste, François Fleuret

This paper proposes a simple yet efficient high-altitude wind nowcasting pipeline.

GeoNeRF: Generalizing NeRF with Geometry Priors

2 code implementations CVPR 2022 Mohammad Mahdi Johari, Yann Lepoittevin, François Fleuret

To render a novel view, the geometry reasoner first constructs cascaded cost volumes for each nearby source view.

Neural Rendering Novel View Synthesis

Test time Adaptation through Perturbation Robustness

1 code implementation19 Oct 2021 Prabhu Teja Sivaprasad, François Fleuret

Data samples generated by several real world processes are dynamic in nature \textit{i. e.}, their characteristics vary with time.

Test-time Adaptation Transfer Learning

Speeding up PCA with priming

no code implementations8 Sep 2021 Bálint Máté, François Fleuret

This algorithm first runs any approximate-PCA method to get an initial estimate of the principal components (priming), and then applies an exact PCA in the subspace they span.

Language Models are Few-Shot Butlers

1 code implementation EMNLP 2021 Vincent Micheli, François Fleuret

Pretrained language models demonstrate strong performance in most NLP tasks when fine-tuned on small task-specific datasets.

reinforcement-learning Reinforcement Learning (RL)

Unsupervised clustering of series using dynamic programming and neural processes

no code implementations26 Jan 2021 Karthigan Sinnathamby, Chang-Yu Hou, Lalitha Venkataramanan, Vasileios-Marios Gkortsas, François Fleuret

Following the work of arXiv:2101. 09512, we are interested in clustering a given multi-variate series in an unsupervised manner.

Clustering

Computing Preimages of Deep Neural Networks with Applications to Safety

no code implementations1 Jan 2021 Kyle Matoba, François Fleuret

To apply an algorithm in a sensitive domain it is important to understand the set of input values that result in specific decisions.

Collision Avoidance Decision Making

Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention

6 code implementations ICML 2020 Angelos Katharopoulos, Apoorv Vyas, Nikolaos Pappas, François Fleuret

Transformers achieve remarkable performance in several tasks but due to their quadratic complexity, with respect to the input's length, they are prohibitively slow for very long sequences.

D4RL Language Modelling +1

Multi-task Reinforcement Learning with a Planning Quasi-Metric

no code implementations8 Feb 2020 Vincent Micheli, Karthigan Sinnathamby, François Fleuret

We introduce a new reinforcement learning approach combining a planning quasi-metric (PQM) that estimates the number of steps required to go from any state to another, with task-specific "aimers" that compute a target state to reach a given goal.

reinforcement-learning Reinforcement Learning (RL)

Optimizer Benchmarking Needs to Account for Hyperparameter Tuning

no code implementations ICML 2020 Prabhu Teja Sivaprasad, Florian Mai, Thijs Vogels, Martin Jaggi, François Fleuret

The performance of optimizers, particularly in deep learning, depends considerably on their chosen hyperparameter configuration.

Benchmarking

Processing Megapixel Images with Deep Attention-Sampling Models

2 code implementations3 May 2019 Angelos Katharopoulos, François Fleuret

We show that sampling from the attention distribution results in an unbiased estimator of the full model with minimal variance, and we derive an unbiased estimator of the gradient that we use to train our model end-to-end with a normal SGD procedure.

Deep Attention

Reducing Noise in GAN Training with Variance Reduced Extragradient

no code implementations NeurIPS 2019 Tatjana Chavdarova, Gauthier Gidel, François Fleuret, Simon Lacoste-Julien

We study the effect of the stochastic gradient noise on the training of generative adversarial networks (GANs) and show that it can prevent the convergence of standard game optimization methods, while the batch version converges.

Not All Samples Are Created Equal: Deep Learning with Importance Sampling

2 code implementations ICML 2018 Angelos Katharopoulos, François Fleuret

Deep neural network training spends most of the computation on examples that are properly handled, and could be ignored.

Image Classification

SGAN: An Alternative Training of Generative Adversarial Networks

no code implementations CVPR 2018 Tatjana Chavdarova, François Fleuret

The Generative Adversarial Networks (GANs) have demonstrated impressive performance for data synthesis, and are now used in a wide range of computer vision tasks.

Biased Importance Sampling for Deep Neural Network Training

1 code implementation31 May 2017 Angelos Katharopoulos, François Fleuret

Importance sampling has been successfully used to accelerate stochastic optimization in many convex problems.

Image Classification Language Modelling +1

Deep Multi-camera People Detection

no code implementations15 Feb 2017 Tatjana Chavdarova, François Fleuret

The former does not exploit joint information, whereas the latter deals with ambiguous input due to the foreground blobs becoming more and more interconnected as the number of targets increases.

Predicting the dynamics of 2d objects with a deep residual network

no code implementations13 Oct 2016 François Fleuret

We investigate how a residual network can learn to predict the dynamics of interacting shapes purely as an image-to-image regression task.

K-Medoids For K-Means Seeding

1 code implementation NeurIPS 2017 James Newling, François Fleuret

We run experiments showing that algorithm clarans (Ng et al., 2005) finds better K-medoids solutions than the Voronoi iteration algorithm.

Data Structures and Algorithms

A Sub-Quadratic Exact Medoid Algorithm

no code implementations23 May 2016 James Newling, François Fleuret

We present a new algorithm, trimed, for obtaining the medoid of a set, that is the element of the set which minimises the mean distance to all other elements.

Nested Mini-Batch K-Means

1 code implementation NeurIPS 2016 James Newling, François Fleuret

A new algorithm is proposed which accelerates the mini-batch k-means algorithm of Sculley (2010) by using the distance bounding approach of Elkan (2003).

Fast K-Means with Accurate Bounds

2 code implementations8 Feb 2016 James Newling, François Fleuret

We propose a novel accelerated exact k-means algorithm, which performs better than the current state-of-the-art low-dimensional algorithm in 18 of 22 experiments, running up to 3 times faster.

Clustering Vector Quantization (k-means problem)

Kullback-Leibler Proximal Variational Inference

no code implementations NeurIPS 2015 Mohammad E. Khan, Pierre Baque, François Fleuret, Pascal Fua

Secondly, we use the proximal framework to derive efficient variational algorithms for non-conjugate models.

Variational Inference

A provably convergent alternating minimization method for mean field inference

no code implementations20 Feb 2015 Pierre Baqué, Jean-Hubert Hours, François Fleuret, Pascal Fua

Mean-Field is an efficient way to approximate a posterior distribution in complex graphical models and constitutes the most popular class of Bayesian variational approximation methods.

Reservoir Boosting : Between Online and Offline Ensemble Learning

no code implementations NeurIPS 2013 Leonidas Lefakis, François Fleuret

We propose to train an ensemble with the help of a reservoir in which the learning algorithm can store a limited number of samples.

Ensemble Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.