no code implementations • ICML 2020 • Fariborz Salehi, Ehsan Abbasi, Babak Hassibi
The performance of hard-margin SVM has been recently analyzed in~\cite{montanari2019generalization, deng2019model}.
no code implementations • 27 Oct 2022 • Taylan Kargin, Fariborz Salehi, Babak Hassibi
The stochastic mirror descent (SMD) algorithm is a general class of training algorithms, which includes the celebrated stochastic gradient descent (SGD), as a special case.
no code implementations • 29 Oct 2020 • Fariborz Salehi, Babak Hassibi
To this end, in this paper we consider the problem of binary classification with adversarial perturbations.
no code implementations • 29 Oct 2020 • Fariborz Salehi, Ehsan Abbasi, Babak Hassibi
We also provide a detailed study for three special cases: ($1$) $\ell_2$-GMM that is the max-margin classifier, ($2$) $\ell_1$-GMM which encourages sparsity, and ($3$) $\ell_{\infty}$-GMM which is often used when the parameter vector has binary entries.
no code implementations • 30 Nov 2019 • Keith Bonawitz, Fariborz Salehi, Jakub Konečný, Brendan Mcmahan, Marco Gruteser
Federated Learning enables mobile devices to collaboratively learn a shared inference model while keeping all the training data on a user's device, decoupling the ability to do machine learning from the need to store the data in the cloud.
no code implementations • NeurIPS 2019 • Fariborz Salehi, Ehsan Abbasi, Babak Hassibi
In both cases, we obtain explicit expressions for various performance metrics and can find the values of the regularizer parameter that optimizes the desired performance.
no code implementations • NeurIPS 2018 • Fariborz Salehi, Ehsan Abbasi, Babak Hassibi
The problem of estimating an unknown signal, $\mathbf x_0\in \mathbb R^n$, from a vector $\mathbf y\in \mathbb R^m$ consisting of $m$ magnitude-only measurements of the form $y_i=|\mathbf a_i\mathbf x_0|$, where $\mathbf a_i$'s are the rows of a known measurement matrix $\mathbf A$ is a classical problem known as phase retrieval.