Search Results for author: Marco Cuturi

Found 76 papers, 32 papers with code

Supervised Quantile Normalization for Low Rank Matrix Factorization

no code implementations ICML 2020 Marco Cuturi, Olivier Teboul, Jonathan Niles-Weed, Jean-Philippe Vert

Low rank matrix factorization is a fundamental building block in machine learning, used for instance to summarize gene expression profile data or word-document counts.

On a Neural Implementation of Brenier's Polar Factorization

no code implementations5 Mar 2024 Nina Vesseron, Marco Cuturi

In 1991, Brenier proved a theorem that generalizes the $QR$ decomposition for square matrices -- factored as PSD $\times$ unitary -- to any vector field $F:\mathbb{R}^d\rightarrow \mathbb{R}^d$.

Careful with that Scalpel: Improving Gradient Surgery with an EMA

no code implementations5 Feb 2024 Yu-Guan Hsieh, James Thornton, Eugene Ndiaye, Michal Klein, Marco Cuturi, Pierre Ablin

Beyond minimizing a single training loss, many deep learning estimation pipelines rely on an auxiliary objective to quantify and encourage desirable properties of the model (e. g. performance on another dataset, robustness, agreement with a prior).

Structured Transforms Across Spaces with Cost-Regularized Optimal Transport

no code implementations9 Nov 2023 Othmane Sebbouh, Marco Cuturi, Gabriel Peyré

Matching a source to a target probability measure is often solved by instantiating a linear optimal transport (OT) problem, parameterized by a ground cost function that quantifies discrepancy between points.

A Specialized Semismooth Newton Method for Kernel-Based Optimal Transport

no code implementations21 Oct 2023 Tianyi Lin, Marco Cuturi, Michael I. Jordan

Kernel-based optimal transport (OT) estimators offer an alternative, functional estimation procedure to address OT problems from samples.

Entropic (Gromov) Wasserstein Flow Matching with GENOT

no code implementations13 Oct 2023 Dominik Klein, Théo Uscidda, Fabian Theis, Marco Cuturi

Optimal transport (OT) theory has reshaped the field of generative modeling: Combined with neural networks, recent \textit{Neural OT} (N-OT) solvers use OT as an inductive bias, to focus on ``thrifty'' mappings that minimize average displacement costs.

Inductive Bias

Simulation-based Inference for Cardiovascular Models

no code implementations26 Jul 2023 Antoine Wehenkel, Jens Behrmann, Andrew C. Miller, Guillermo Sapiro, Ozan Sener, Marco Cuturi, Jörn-Henrik Jacobsen

Over the past decades, hemodynamics simulators have steadily evolved and have become tools of choice for studying cardiovascular systems in-silico.

Learning Costs for Structured Monge Displacements

no code implementations20 Jun 2023 Michal Klein, Aram-Alexandre Pooladian, Pierre Ablin, Eugène Ndiaye, Jonathan Niles-Weed, Marco Cuturi

Because of such difficulties, existing approaches rarely depart from the default choice of estimating such maps with the simple squared-Euclidean distance as the ground cost, $c(x, y)=\|x-y\|^2_2$.

The Monge Gap: A Regularizer to Learn All Transport Maps

no code implementations9 Feb 2023 Théo Uscidda, Marco Cuturi

That gap quantifies how far a map $T$ deviates from the ideal properties we expect from a $c$-OT map.

MORPH

Monge, Bregman and Occam: Interpretable Optimal Transport in High-Dimensions with Feature-Sparse Maps

no code implementations8 Feb 2023 Marco Cuturi, Michal Klein, Pierre Ablin

Optimal transport (OT) theory focuses, among all maps $T:\mathbb{R}^d\rightarrow \mathbb{R}^d$ that can morph a probability measure onto another, on those that are the ``thriftiest'', i. e. such that the averaged cost $c(x, T(x))$ between $x$ and its image $T(x)$ be as small as possible.

Dimensionality Reduction MORPH

Supervised Training of Conditional Monge Maps

1 code implementation28 Jun 2022 Charlotte Bunne, Andreas Krause, Marco Cuturi

To account for that context in OT estimation, we introduce CondOT, a multi-task approach to estimate a family of OT maps conditioned on a context variable, using several pairs of measures $\left(\mu_i, \nu_i\right)$ tagged with a context label $c_i$.

Rethinking Initialization of the Sinkhorn Algorithm

no code implementations15 Jun 2022 James Thornton, Marco Cuturi

While the optimal transport (OT) problem was originally formulated as a linear program, the addition of entropic regularization has proven beneficial both computationally and statistically, for many applications.

Low-rank Optimal Transport: Approximation, Statistics and Debiasing

no code implementations24 May 2022 Meyer Scetbon, Marco Cuturi

The matching principles behind optimal transport (OT) play an increasingly important role in machine learning, a trend which can be observed when OT is used to disambiguate datasets in applications (e. g. single-cell genomics) or used to improve more complex methods (e. g. balanced attention in transformers or self-supervised learning).

Self-Supervised Learning

Simultaneous Multiple-Prompt Guided Generation Using Differentiable Optimal Transport

no code implementations18 Apr 2022 Yingtao Tian, Marco Cuturi, David Ha

Recent advances in deep learning, such as powerful generative models and joint text-image embeddings, have provided the computational creativity community with new tools, opening new perspectives for artistic pursuits.

Image Generation

Averaging Spatio-temporal Signals using Optimal Transport and Soft Alignments

1 code implementation11 Mar 2022 Hicham Janati, Marco Cuturi, Alexandre Gramfort

These complex datasets, describing dynamics with both time and spatial components, pose new challenges for data analysis.

Dynamic Time Warping

The Schrödinger Bridge between Gaussian Measures has a Closed Form

no code implementations11 Feb 2022 Charlotte Bunne, Ya-Ping Hsieh, Marco Cuturi, Andreas Krause

The static optimal transport $(\mathrm{OT})$ problem between Gaussians seeks to recover an optimal map, or more generally a coupling, to morph a Gaussian into another.

Gaussian Processes MORPH

Optimal Transport Tools (OTT): A JAX Toolbox for all things Wasserstein

1 code implementation28 Jan 2022 Marco Cuturi, Laetitia Meng-Papaxanthos, Yingtao Tian, Charlotte Bunne, Geoff Davis, Olivier Teboul

Optimal transport tools (OTT-JAX) is a Python toolbox that can solve optimal transport problems between point clouds and histograms.

Randomized Stochastic Gradient Descent Ascent

no code implementations25 Nov 2021 Othmane Sebbouh, Marco Cuturi, Gabriel Peyré

RSGDA can be parameterized using optimal loop sizes that guarantee the best convergence rates known to hold for SGDA.

Proximal Optimal Transport Modeling of Population Dynamics

1 code implementation11 Jun 2021 Charlotte Bunne, Laetitia Meng-Papaxanthos, Andreas Krause, Marco Cuturi

We propose to model these trajectories as collective realizations of a causal Jordan-Kinderlehrer-Otto (JKO) flow of measures: The JKO scheme posits that the new configuration taken by a population at time $t+1$ is one that trades off an improvement, in the sense that it decreases an energy, while remaining close (in Wasserstein distance) to the previous configuration observed at $t$.

Linear-Time Gromov Wasserstein Distances using Low Rank Couplings and Costs

1 code implementation NeurIPS 2021 Meyer Scetbon, Gabriel Peyré, Marco Cuturi

The ability to align points across two related yet incomparable point clouds (e. g. living in different spaces) plays an important role in machine learning.

Efficient and Modular Implicit Differentiation

1 code implementation NeurIPS 2021 Mathieu Blondel, Quentin Berthet, Marco Cuturi, Roy Frostig, Stephan Hoyer, Felipe Llinares-López, Fabian Pedregosa, Jean-Philippe Vert

In this paper, we propose automatic implicit differentiation, an efficient and modular approach for implicit differentiation of optimization problems.

Meta-Learning

Low-Rank Sinkhorn Factorization

1 code implementation8 Mar 2021 Meyer Scetbon, Marco Cuturi, Gabriel Peyré

Because matrix-vector products are pervasive in the Sinkhorn algorithm, several works have proposed to \textit{approximate} kernel matrices appearing in its iterations using low-rank factors.

Entropic Optimal Transport between Unbalanced Gaussian Measures has a Closed Form

no code implementations NeurIPS 2020 Hicham Janati, Boris Muzellec, Gabriel Peyré, Marco Cuturi

Although optimal transport (OT) problems admit closed form solutions in a very few notable cases, e. g. in 1D or between Gaussians, these closed forms have proved extremely fecund for practitioners to define tools inspired from the OT geometry.

Learning with Differentiable Pertubed Optimizers

no code implementations NeurIPS 2020 Quentin Berthet, Mathieu Blondel, Olivier Teboul, Marco Cuturi, Jean-Philippe Vert, Francis Bach

Machine learning pipelines often rely on optimizers procedures to make discrete decisions (e. g., sorting, picking closest neighbors, or shortest paths).

Structured Prediction

On Projection Robust Optimal Transport: Sample Complexity and Model Misspecification

no code implementations22 Jun 2020 Tianyi Lin, Zeyu Zheng, Elynn Y. Chen, Marco Cuturi, Michael. I. Jordan

Yet, the behavior of minimum Wasserstein estimators is poorly understood, notably in high-dimensional regimes or under model misspecification.

Projection Robust Wasserstein Distance and Riemannian Optimization

no code implementations NeurIPS 2020 Tianyi Lin, Chenyou Fan, Nhat Ho, Marco Cuturi, Michael. I. Jordan

Projection robust Wasserstein (PRW) distance, or Wasserstein projection pursuit (WPP), is a robust variant of the Wasserstein distance.

Riemannian optimization

Linear Time Sinkhorn Divergences using Positive Features

1 code implementation NeurIPS 2020 Meyer Scetbon, Marco Cuturi

Although Sinkhorn divergences are now routinely used in data sciences to compare probability distributions, the computational effort required to compute them remains expensive, growing in general quadratically in the size $n$ of the support of these distributions.

Equitable and Optimal Transport with Multiple Agents

no code implementations12 Jun 2020 Meyer Scetbon, Laurent Meunier, Jamal Atif, Marco Cuturi

When there is only one agent, we recover the Optimal Transport problem.

Entropic Optimal Transport between (Unbalanced) Gaussian Measures has a Closed Form

1 code implementation NeurIPS 2020 Hicham Janati, Boris Muzellec, Gabriel Peyré, Marco Cuturi

Although optimal transport (OT) problems admit closed form solutions in a very few notable cases, e. g. in 1D or between Gaussians, these closed forms have proved extremely fecund for practitioners to define tools inspired from the OT geometry.

Statistics Theory Statistics Theory

Debiased Sinkhorn barycenters

3 code implementations ICML 2020 Hicham Janati, Marco Cuturi, Alexandre Gramfort

However, entropy brings some inherent smoothing bias, resulting for example in blurred barycenters.

Noisy Adaptive Group Testing using Bayesian Sequential Experimental Design

1 code implementation26 Apr 2020 Marco Cuturi, Olivier Teboul, Quentin Berthet, Arnaud Doucet, Jean-Philippe Vert

Our goal in this paper is to propose new group testing algorithms that can operate in a noisy setting (tests can be mistaken) to decide adaptively (looking at past results) which groups to test next, with the goal to converge to a good detection, as quickly, and with as few tests as possible.

Experimental Design

Learning with Differentiable Perturbed Optimizers

2 code implementations20 Feb 2020 Quentin Berthet, Mathieu Blondel, Olivier Teboul, Marco Cuturi, Jean-Philippe Vert, Francis Bach

Machine learning pipelines often rely on optimization procedures to make discrete decisions (e. g., sorting, picking closest neighbors, or shortest paths).

Structured Prediction

Fixed-Support Wasserstein Barycenters: Computational Hardness and Fast Algorithm

no code implementations NeurIPS 2020 Tianyi Lin, Nhat Ho, Xi Chen, Marco Cuturi, Michael. I. Jordan

We study the fixed-support Wasserstein barycenter problem (FS-WBP), which consists in computing the Wasserstein barycenter of $m$ discrete probability measures supported on a finite metric space of size $n$.

Open-Ended Question Answering

Missing Data Imputation using Optimal Transport

1 code implementation ICML 2020 Boris Muzellec, Julie Josse, Claire Boyer, Marco Cuturi

Missing data is a crucial issue when applying machine learning algorithms to real-world datasets.

Imputation

Regularized Optimal Transport is Ground Cost Adversarial

no code implementations ICML 2020 François-Pierre Paty, Marco Cuturi

In this work we depart from this practical perspective and propose a new interpretation of regularization as a robust mechanism, and show using Fenchel duality that any convex regularization of OT can be interpreted as ground cost adversarial.

Supervised Quantile Normalization for Low-rank Matrix Approximation

no code implementations8 Feb 2020 Marco Cuturi, Olivier Teboul, Jonathan Niles-Weed, Jean-Philippe Vert

Low rank matrix factorization is a fundamental building block in machine learning, used for instance to summarize gene expression profile data or word-document counts.

Fast and Robust Comparison of Probability Measures in Heterogeneous Spaces

1 code implementation5 Feb 2020 Ryoma Sato, Marco Cuturi, Makoto Yamada, Hisashi Kashima

Building on \cite{memoli-2011}, who proposed to represent each point in each distribution as the 1D distribution of its distances to all other points, we introduce in this paper the Anchor Energy (AE) and Anchor Wasserstein (AW) distances, which are respectively the energy and Wasserstein distances instantiated on such representations.

Graph Matching Word Embeddings

Differentiable Ranking and Sorting using Optimal Transport

1 code implementation NeurIPS 2019 Marco Cuturi, Olivier Teboul, Jean-Philippe Vert

From this observation, we propose extended rank and sort operators by considering optimal transport (OT) problems (the natural relaxation for assignments) where the auxiliary measure can be any weighted measure supported on $m$ increasing values, where $m \ne n$.

Ground Metric Learning on Graphs

1 code implementation8 Nov 2019 Matthieu Heitz, Nicolas Bonneel, David Coeurjolly, Marco Cuturi, Gabriel Peyré

Optimal transport (OT) distances between probability distributions are parameterized by the ground metric they use between observations.

Metric Learning

Spatio-Temporal Alignments: Optimal transport through space and time

2 code implementations9 Oct 2019 Hicham Janati, Marco Cuturi, Alexandre Gramfort

In this paper, we propose Spatio-Temporal Alignments (STA), a new differentiable formulation of DTW, in which spatial differences between time samples are accounted for using regularized optimal transport (OT).

Dynamic Time Warping Time Series +1

Multi-subject MEG/EEG source imaging with sparse multi-task regression

no code implementations3 Oct 2019 Hicham Janati, Thomas Bazeille, Bertrand Thirion, Marco Cuturi, Alexandre Gramfort

Magnetoencephalography and electroencephalography (M/EEG) are non-invasive modalities that measure the weak electromagnetic fields generated by neural activity.

EEG regression +1

On the Complexity of Approximating Multimarginal Optimal Transport

no code implementations30 Sep 2019 Tianyi Lin, Nhat Ho, Marco Cuturi, Michael. I. Jordan

This provides a first \textit{near-linear time} complexity bound guarantee for approximating the MOT problem and matches the best known complexity bound for the Sinkhorn algorithm in the classical OT setting when $m = 2$.

Differentiable Ranks and Sorting using Optimal Transport

no code implementations28 May 2019 Marco Cuturi, Olivier Teboul, Jean-Philippe Vert

Sorting an array is a fundamental routine in machine learning, one that is used to compute rank-based statistics, cumulative distribution functions (CDFs), quantiles, or to select closest neighbors and labels.

Regularity as Regularization: Smooth and Strongly Convex Brenier Potentials in Optimal Transport

no code implementations26 May 2019 François-Pierre Paty, Alexandre d'Aspremont, Marco Cuturi

On the other hand, one of the greatest achievements of the OT literature in recent years lies in regularity theory: Caffarelli showed that the OT map between two well behaved measures is Lipschitz, or equivalently when considering 2-Wasserstein distances, that Brenier convex potentials (whose gradient yields an optimal map) are smooth.

Domain Adaptation

Precision-Recall Curves Using Information Divergence Frontiers

no code implementations26 May 2019 Josip Djolonga, Mario Lucic, Marco Cuturi, Olivier Bachem, Olivier Bousquet, Sylvain Gelly

Despite the tremendous progress in the estimation of generative models, the development of tools for diagnosing their failures and assessing their performance has advanced at a much slower pace.

Image Generation Information Retrieval +1

Subspace Detours: Building Transport Plans that are Optimal on Subspace Projections

1 code implementation NeurIPS 2019 Boris Muzellec, Marco Cuturi

A popular approach to avoid this curse is to project input measures on lower-dimensional subspaces (1D lines in the case of sliced Wasserstein distances), solve the OT problem between these reduced measures, and settle for the Wasserstein distance between these reductions, rather than that between the original measures.

Domain Adaptation Word Embeddings

Group level MEG/EEG source imaging via optimal transport: minimum Wasserstein estimates

no code implementations13 Feb 2019 Hicham Janati, Thomas Bazeille, Bertrand Thirion, Marco Cuturi, Alexandre Gramfort

Inferring the location of the current sources that generated these magnetic fields is an ill-posed inverse problem known as source imaging.

EEG

Tree-Sliced Variants of Wasserstein Distances

2 code implementations NeurIPS 2019 Tam Le, Makoto Yamada, Kenji Fukumizu, Marco Cuturi

Optimal transport (\OT) theory defines a powerful set of tools to compare probability distributions.

Subspace Robust Wasserstein Distances

no code implementations25 Jan 2019 François-Pierre Paty, Marco Cuturi

Making sense of Wasserstein distances between discrete measures in high-dimensional settings remains a challenge.

Quantization

Stochastic Deep Networks

1 code implementation19 Nov 2018 Gwendoline de Bie, Gabriel Peyré, Marco Cuturi

This allows to design discriminative networks (to classify or reduce the dimensionality of input measures), generative architectures (to synthesize measures) and recurrent pipelines (to predict measure dynamics).

Semi-dual Regularized Optimal Transport

no code implementations13 Nov 2018 Marco Cuturi, Gabriel Peyré

Variational problems that involve Wasserstein distances and more generally optimal transport (OT) theory are playing an increasingly important role in data sciences.

Wasserstein regularization for sparse multi-task regression

1 code implementation20 May 2018 Hicham Janati, Marco Cuturi, Alexandre Gramfort

We argue in this paper that these techniques fail to leverage the spatial information associated to regressors.

regression

Generalizing Point Embeddings using the Wasserstein Space of Elliptical Distributions

2 code implementations NeurIPS 2018 Boris Muzellec, Marco Cuturi

We propose in this work an extension of that approach, which consists in embedding objects as elliptical probability distributions, namely distributions whose densities have elliptical level sets.

valid

Computational Optimal Transport

5 code implementations1 Mar 2018 Gabriel Peyré, Marco Cuturi

Optimal transport (OT) theory can be informally described using the words of the French mathematician Gaspard Monge (1746-1818): A worker with a shovel in hand has to move a large pile of sand lying on a construction site.

MORPH

Sliced Wasserstein Kernel for Persistence Diagrams

no code implementations ICML 2017 Mathieu Carrière, Marco Cuturi, Steve Oudot

To incorporate PDs in a learning pipeline, several kernels have been proposed for PDs with a strong emphasis on the stability of the RKHS distance w. r. t.

Graph Classification Topological Data Analysis

GAN and VAE from an Optimal Transport Point of View

no code implementations6 Jun 2017 Aude Genevay, Gabriel Peyré, Marco Cuturi

This short article revisits some of the ideas introduced in arXiv:1701. 07875 and arXiv:1705. 07642 in a simple setup.

Learning Generative Models with Sinkhorn Divergences

2 code implementations1 Jun 2017 Aude Genevay, Gabriel Peyré, Marco Cuturi

The ability to compare two degenerate probability distributions (i. e. two probability distributions supported on two distinct low-dimensional manifolds living in a much higher-dimensional space) is a crucial problem arising in the estimation of generative models for high-dimensional observations such as those arising in computer vision or natural language.

Soft-DTW: a Differentiable Loss Function for Time-Series

8 code implementations ICML 2017 Marco Cuturi, Mathieu Blondel

We propose in this paper a differentiable learning loss between time series, building upon the celebrated dynamic time warping (DTW) discrepancy.

Dynamic Time Warping Time Series +1

Wasserstein Training of Restricted Boltzmann Machines

no code implementations NeurIPS 2016 Grégoire Montavon, Klaus-Robert Müller, Marco Cuturi

This metric between observations can then be used to define the Wasserstein distance between the distribution induced by the Boltzmann machine on the one hand, and that given by the training sample on the other hand.

Denoising

Wasserstein Discriminant Analysis

1 code implementation29 Aug 2016 Rémi Flamary, Marco Cuturi, Nicolas Courty, Alain Rakotomamonjy

Wasserstein Discriminant Analysis (WDA) is a new supervised method that can improve classification of high-dimensional data by computing a suitable linear map onto a lower dimensional subspace.

Stochastic Optimization for Large-scale Optimal Transport

no code implementations NeurIPS 2016 Genevay Aude, Marco Cuturi, Gabriel Peyré, Francis Bach

We instantiate these ideas in three different setups: (i) when comparing a discrete distribution to another, we show that incremental stochastic optimization schemes can beat Sinkhorn's algorithm, the current state-of-the-art finite dimensional OT solver; (ii) when comparing a discrete distribution to a continuous density, a semi-discrete reformulation of the dual program is amenable to averaged stochastic gradient descent, leading to better performance than approximately solving the problem by discretization ; (iii) when dealing with two continuous densities, we propose a stochastic gradient descent over a reproducing kernel Hilbert space (RKHS).

Stochastic Optimization

On Wasserstein Two Sample Testing and Related Families of Nonparametric Tests

1 code implementation8 Sep 2015 Aaditya Ramdas, Nicolas Garcia, Marco Cuturi

In this work, our central object is the Wasserstein distance, as we form a chain of connections from univariate methods like the Kolmogorov-Smirnov test, PP/QQ plots and ROC/ODC curves, to multivariate tests involving energy statistics and kernel based maximum mean discrepancy.

Two-sample testing Vocal Bursts Valence Prediction

Wasserstein Training of Boltzmann Machines

no code implementations7 Jul 2015 Grégoire Montavon, Klaus-Robert Müller, Marco Cuturi

The Boltzmann machine provides a useful framework to learn highly complex, multimodal and multiscale data distributions that occur in the real world.

Denoising

Principal Geodesic Analysis for Probability Measures under the Optimal Transport Metric

no code implementations NeurIPS 2015 Vivien Seguy, Marco Cuturi

Given a family of probability measures in P(X), the space of probability measures on a Hilbert space X, our goal in this paper is to highlight one ore more curves in P(X) that summarize efficiently that family.

Fast Optimal Transport Averaging of Neuroimaging Data

no code implementations30 Mar 2015 Alexandre Gramfort, Gabriel Peyré, Marco Cuturi

Data are large, the geometry of the brain is complex and the between subjects variability leads to spatially or temporally non-overlapping effects of interest.

A Smoothed Dual Approach for Variational Wasserstein Problems

1 code implementation9 Mar 2015 Marco Cuturi, Gabriel Peyré

Variational problems that involve Wasserstein distances have been recently proposed to summarize and learn from probability measures.

Iterative Bregman Projections for Regularized Transportation Problems

1 code implementation16 Dec 2014 Jean-David Benamou, Guillaume Carlier, Marco Cuturi, Luca Nenna, Gabriel Peyré

This article details a general numerical framework to approximate so-lutions to linear programs related to optimal transport.

Numerical Analysis Analysis of PDEs

Sinkhorn Distances: Lightspeed Computation of Optimal Transport

no code implementations NeurIPS 2013 Marco Cuturi

Optimal transportation distances are a fundamental family of parameterized distances for histograms in the probability simplex.

Fast Computation of Wasserstein Barycenters

2 code implementations16 Oct 2013 Marco Cuturi, Arnaud Doucet

We present new algorithms to compute the mean of a set of empirical probability measures under the optimal transport metric.

Constrained Clustering

Sinkhorn Distances: Lightspeed Computation of Optimal Transportation Distances

10 code implementations NeurIPS 2013 Marco Cuturi

Optimal transportation distances are a fundamental family of parameterized distances for histograms.

Retrieval

White Functionals for Anomaly Detection in Dynamical Systems

no code implementations NeurIPS 2009 Marco Cuturi, Jean-Philippe Vert, Alexandre d'Aspremont

The candidate functionals are estimated in a subset of a reproducing kernel Hilbert space associated with the set where the process takes values.

Anomaly Detection

Cannot find the paper you are looking for? You can Submit a new open access paper.