no code implementations • 22 Oct 2022 • Huan He, Shifan Zhao, Ziyuan Tang, Joyce C Ho, Yousef Saad, Yuanzhe Xi
Nonlinear acceleration methods are powerful techniques to speed up fixed-point iterations.
1 code implementation • 6 Oct 2021 • Huan He, Shifan Zhao, Yuanzhe Xi, Joyce C Ho, Yousef Saad
We also empirically show that GDA-AMsolves a variety of minimax problems and improves GAN training on several datasets
no code implementations • ICLR 2022 • Huan He, Shifan Zhao, Yuanzhe Xi, Joyce Ho, Yousef Saad
We also empirically show that GDA-AM solves a variety of minimax problems and improves GAN training on several datasets
no code implementations • 22 Jun 2021 • Jie Chen, Yousef Saad, Zechen Zhang
The general method of graph coarsening or graph reduction has been a remarkably useful and ubiquitous tool in scientific computing and it is now just starting to have a similar impact in machine learning.
2 code implementations • 25 Jun 2019 • Jia Shi, Ruipeng Li, Yuanzhe Xi, Yousef Saad, Maarten V. de Hoop
A Continuous Galerkin method-based approach is presented to compute the seismic normal modes of rotating planets.
Computational Physics Earth and Planetary Astrophysics Geophysics 86-08, 86-04, 85-04, 85-08, 85-10, 15A18, 65N25, 65N30
no code implementations • 8 Oct 2018 • Shashanka Ubaru, Abd-Krim Seghouane, Yousef Saad
In this paper, we consider the problem of simultaneously estimating the dimension of the principal (dominant) subspace of these covariance matrices and obtaining an approximation to the subspace.
1 code implementation • 14 Feb 2018 • Ruipeng Li, Yuanzhe Xi, Lucas Erlandson, Yousef Saad
This paper describes a software package called EVSL (for EigenValues Slicing Library) for solving large sparse real symmetric standard and generalized eigenvalue problems.
Numerical Analysis
no code implementations • NeurIPS 2017 • Gang Wang, Georgios Giannakis, Yousef Saad, Jie Chen
For certain random measurement models, the proposed procedure returns the true solution $\bm{x}$ with high probability in time proportional to reading the data $\{(\bm{a}_i;y_i)\}_{1\le i \le m}$, provided that the number $m$ of equations is some constant $c>0$ times the number $n$ of unknowns, that is, $m\ge cn$.
no code implementations • 1 Nov 2017 • Shashanka Ubaru, Yousef Saad
To tackle these problems, we consider algorithms that are based primarily on coarsening techniques, possibly combined with random sampling.
no code implementations • 29 May 2017 • Gang Wang, Georgios B. Giannakis, Yousef Saad, Jie Chen
This paper deals with finding an $n$-dimensional solution $x$ to a system of quadratic equations of the form $y_i=|\langle{a}_i, x\rangle|^2$ for $1\le i \le m$, which is also known as phase retrieval and is NP-hard in general.
1 code implementation • 15 May 2017 • Difeng Cai, Edmond Chow, Yousef Saad, Yuanzhe Xi
This paper presents an efficient method to perform Structured Matrix Approximation by Separation and Hierarchy (SMASH), when the original dense matrix is associated with a kernel function.
Numerical Analysis
no code implementations • 28 Feb 2017 • Raphael Petegrosso, Wei zhang, Zhuliu Li, Yousef Saad, Rui Kuang
The success of semi-supervised learning crucially relies on the scalability to a huge amount of unlabelled data that are needed to capture the underlying manifold structure for better classification.
no code implementations • 19 Aug 2016 • Shashanka Ubaru, Yousef Saad, Abd-Krim Seghouane
In this paper, we present two computationally inexpensive techniques to estimate the approximate ranks of such large matrices.
no code implementations • 30 Dec 2015 • Shashanka Ubaru, Arya Mazumdar, Yousef Saad
In this paper, we show how matrices from error correcting codes can be used to find such low rank approximations and matrix decompositions, and extend the framework to linear least squares regression problems.
no code implementations • NeurIPS 2012 • Thanh Ngo, Yousef Saad
This paper describes gradient methods based on a scaled metric on the Grassmann manifold for low-rank matrix completion.