1 code implementation • 6 Nov 2024 • Mauricio Velasco, Kaiying O'Hare, Bernardo Rychtenberg, Soledad Villar
Our theoretical results extend well-known transferability theorems for GNNs to the case of several simultaneous graphs (GtNNs) and provide a strict improvement on what is currently known even in the GNN case.
no code implementations • 28 Aug 2024 • George A. Kevrekidis, Mauro Maggioni, Soledad Villar, Yannis G. Kevrekidis
Conformal Autoencoders are a neural network architecture that imposes orthogonality conditions between the gradients of latent variables towards achieving disentangled representations of data.
1 code implementation • 3 Jun 2024 • Wilson G. Gregory, Josué Tonelli-Cueto, Nicholas F. Marshall, Andrew S. Lee, Soledad Villar
Our numerical results show that the proposed equivariant machine learning models can learn spectral methods that outperform the best theoretically known spectral methods in some regimes.
1 code implementation • 28 May 2024 • David W. Hogg, Soledad Villar
The question in the title is being asked of all of the natural sciences; that is, we are calling on the scientific communities to take a step back and consider the role and value of ML in their fields; the (partial) answers we give here come from the particular perspective of physics.
no code implementations • 13 May 2024 • Ben Blum-Smith, Ningyuan Huang, Marco Cuturi, Soledad Villar
In this work, we present a mathematical formulation for machine learning of (1) functions on symmetric matrices that are invariant with respect to the action of permutations by conjugation, and (2) functions on point clouds that are invariant with respect to rotations, reflections, and permutations of the points.
1 code implementation • NeurIPS 2023 • Ningyuan Huang, Ron Levie, Soledad Villar
However, these two symmetries are fundamentally different: The translation equivariance of CNNs corresponds to symmetries of the fixed domain acting on the image signals (sometimes known as active symmetries), whereas in GNNs any permutation acts on both the graph signals and the graph domain (sometimes described as passive symmetries).
1 code implementation • 24 Jun 2023 • Sharut Gupta, Joshua Robinson, Derek Lim, Soledad Villar, Stefanie Jegelka
Specifically, in the contrastive learning setting, we introduce an equivariance objective and theoretically prove that its minima forces augmentations on input space to correspond to rotations on the spherical embedding space.
1 code implementation • NeurIPS 2023 • Jan Böker, Ron Levie, Ningyuan Huang, Soledad Villar, Christopher Morris
In particular, we characterize the expressive power of MPNNs in terms of the tree distance, which is a graph distance based on the concept of fractional isomorphisms, and substructure counts via tree homomorphisms, showing that these concepts have the same expressive power as the $1$-WL and MPNNs on graphons.
1 code implementation • 21 May 2023 • Wilson Gregory, David W. Hogg, Ben Blum-Smith, Maria Teresa Arias, Kaze W. K. Wong, Soledad Villar
We use representation theory to quantify the dimension of the space of equivariant polynomial functions on 2-dimensional vector images.
no code implementations • 31 Jan 2023 • Soledad Villar, David W. Hogg, Weichi Yao, George A. Kevrekidis, Bernhard Schölkopf
We discuss links to causal modeling, and argue that the implementation of passive symmetries is particularly valuable when the goal of the learning problem is to generalize out of sample.
1 code implementation • 28 Nov 2022 • Charles Clum, Dustin G. Mixon, Soledad Villar, Kaiying Xie
This lower bound is data-driven; it does not make any assumption on the data nor how it is generated.
1 code implementation • 6 Nov 2022 • Luana Ruiz, Ningyuan Huang, Soledad Villar
In this work we propose a random graph model that can produce graphs at different levels of sparsity.
no code implementations • 26 Oct 2022 • Carey E. Priebe, Ningyuan Huang, Soledad Villar, Cong Mu, Li Chen
We conjecture that for general label noise, mitigation strategies that make use of the noisy data will outperform those that ignore the noisy data.
no code implementations • 30 Sep 2022 • Efe Onaran, Soledad Villar
The shuffled linear regression problem aims to recover linear relationships in datasets where the correspondence between input and output is unknown.
no code implementations • 29 Sep 2022 • Ben Blum-Smith, Soledad Villar
Inspired by constraints from physical law, equivariant machine learning restricts the learning to a hypothesis class where all the functions are equivariant with respect to some group action.
1 code implementation • 24 Sep 2022 • Ningyuan Huang, Soledad Villar, Carey E. Priebe, Da Zheng, Chengyue Huang, Lin Yang, Vladimir Braverman
Graph Neural Networks (GNNs) are powerful deep learning methods for Non-Euclidean data.
no code implementations • 28 Jul 2022 • Nabeel Sarwar, Wilson Gregory, George A Kevrekidis, Soledad Villar, Bianca Dumitrascu
Single-cell RNA-seq data allow the quantification of cell type differences across a growing set of biological contexts.
1 code implementation • 2 Apr 2022 • Soledad Villar, Weichi Yao, David W. Hogg, Ben Blum-Smith, Bianca Dumitrascu
Units equivariance (or units covariance) is the exact symmetry that follows from the requirement that relationships among measured quantities of physics relevance must obey self-consistent dimensional scalings.
no code implementations • 19 Jan 2022 • Ashwin De Silva, Rahul Ramesh, Lyle Ungar, Marshall Hussain Shuler, Noah J. Cowan, Michael Platt, Chen Li, Leyla Isik, Seung-Eon Roh, Adam Charles, Archana Venkataraman, Brian Caffo, Javier J. How, Justus M Kebschull, John W. Krakauer, Maxim Bichuch, Kaleab Alemayehu Kinfu, Eva Yezerets, Dinesh Jayaraman, Jong M. Shin, Soledad Villar, Ian Phillips, Carey E. Priebe, Thomas Hartung, Michael I. Miller, Jayanta Dey, Ningyuan, Huang, Eric Eaton, Ralph Etienne-Cummings, Elizabeth L. Ogburn, Randal Burns, Onyema Osuagwu, Brett Mensh, Alysson R. Muotri, Julia Brown, Chris White, Weiwei Yang, Andrei A. Rusu, Timothy Verstynen, Konrad P. Kording, Pratik Chaudhari, Joshua T. Vogelstein
We conjecture that certain sequences of tasks are not retrospectively learnable (in which the data distribution is fixed), but are prospectively learnable (in which distributions may be dynamic), suggesting that prospective learning is more difficult in kind than retrospective learning.
no code implementations • 18 Jan 2022 • Ningyuan Huang, Soledad Villar
Graph neural networks are designed to learn functions on graphs.
2 code implementations • 7 Oct 2021 • Weichi Yao, Kate Storey-Fisher, David W. Hogg, Soledad Villar
Physical systems obey strict symmetry principles.
2 code implementations • NeurIPS 2021 • Soledad Villar, David W. Hogg, Kate Storey-Fisher, Weichi Yao, Ben Blum-Smith
There has been enormous progress in the last few years in designing neural networks that respect the fundamental symmetries and coordinate freedoms of physical law.
no code implementations • 15 Jan 2021 • David W. Hogg, Soledad Villar
We emphasize that it is often valuable to choose far more parameters than data points, despite folk rules to the contrary: Suitably regularized models with enormous numbers of parameters generalize well and make good predictions for held-out data; over-fitting is not (mainly) a problem of having too many parameters.
1 code implementation • 23 Nov 2020 • Ningyuan Huang, David W. Hogg, Soledad Villar
This realization brought back the study of linear models for regression, including ordinary least squares (OLS), which, like deep learning, shows a "double-descent" behavior: (1) The risk (expected out-of-sample prediction error) can grow arbitrarily when the number of parameters $p$ approaches the number of samples $n$, and (2) the risk decreases with $p$ for $p>n$, sometimes achieving a lower value than the lowest risk for $p<n$.
1 code implementation • NeurIPS 2020 • Zhengdao Chen, Lei Chen, Soledad Villar, Joan Bruna
We also prove positive results for k-WL and k-IGNs as well as negative results for k-WL with a finite number of iterations.
1 code implementation • 6 Jan 2020 • Andrew J. Blumberg, Mathieu Carriere, Michael A. Mandell, Raul Rabadan, Soledad Villar
Comparing and aligning large datasets is a pervasive problem occurring across many different knowledge domains.
no code implementations • 15 Aug 2019 • Weichi Yao, Afonso S. Bandeira, Soledad Villar
In particular we consider Graph Neural Networks (GNNs) -- a class of neural networks designed to learn functions on graphs -- and we apply them to the max-cut problem on random regular graphs.
1 code implementation • NeurIPS 2019 • Zhengdao Chen, Soledad Villar, Lei Chen, Joan Bruna
We further develop a framework of the expressive power of GNNs that incorporates both of these viewpoints using the language of sigma-algebra, through which we compare the expressive power of different types of GNNs together with other graph isomorphism tests.
Ranked #33 on
Graph Regression
on ZINC-500k
1 code implementation • 6 Dec 2018 • Culver McWhirter, Dustin G. Mixon, Soledad Villar
Given labeled points in a high-dimensional vector space, we seek a low-dimensional subspace such that projecting onto this subspace maintains some prescribed distance between points of differing labels.
no code implementations • 25 Mar 2018 • Dustin G. Mixon, Soledad Villar
It has been experimentally established that deep neural networks can be used to produce good generative models for real world data.
no code implementations • 3 Oct 2017 • Dustin G. Mixon, Soledad Villar
Efficient algorithms for $k$-means clustering frequently converge to suboptimal partitions, and given a partition, it is difficult to detect $k$-means optimality.
3 code implementations • 22 Jun 2017 • Alex Nowak, Soledad Villar, Afonso S. Bandeira, Joan Bruna
Inverse problems correspond to a certain type of optimization problems formulated over appropriate input distributions.
no code implementations • 17 Oct 2016 • Soledad Villar, Afonso S. Bandeira, Andrew J. Blumberg, Rachel Ward
The Gromov-Hausdorff distance provides a metric on the set of isometry classes of compact metric spaces.
no code implementations • 22 Feb 2016 • Dustin G. Mixon, Soledad Villar, Rachel Ward
We introduce a model-free relax-and-round algorithm for k-means clustering based on a semidefinite relaxation due to Peng and Wei.
no code implementations • 26 Sep 2015 • Takayuki Iguchi, Dustin G. Mixon, Jesse Peterson, Soledad Villar
First, we prove that Peng and Wei's semidefinite relaxation of k-means is tight with high probability under a distribution of planted clusters called the stochastic ball model.
no code implementations • 18 May 2015 • Takayuki Iguchi, Dustin G. Mixon, Jesse Peterson, Soledad Villar
Recently, Awasthi et al. introduced an SDP relaxation of the $k$-means problem in $\mathbb R^m$.
no code implementations • 18 Aug 2014 • Pranjal Awasthi, Afonso S. Bandeira, Moses Charikar, Ravishankar Krishnaswamy, Soledad Villar, Rachel Ward
Under the same distributional model, the $k$-means LP relaxation fails to recover such clusters at separation as large as $\Delta = 4$.