no code implementations • 3 Dec 2024 • Gregory Dexter, Petros Drineas, Rajiv Khanna
We provide space complexity lower bounds for data structures that approximate logistic loss up to $\epsilon$-relative error on a logistic regression problem with data $\mathbf{X} \in \mathbb{R}^{n \times d}$ and labels $\mathbf{y} \in \{-1, 1\}^d$.
no code implementations • 22 Jan 2024 • Gregory Dexter, Borja Ocejo, Sathiya Keerthi, Aman Gupta, Ayan Acharya, Rajiv Khanna
In this paper, we delve deeper into the relationship between linear stability and sharpness.
no code implementations • 30 Sep 2023 • Young In Kim, Pratiksha Agrawal, Johannes O. Royset, Rajiv Khanna
In this work, we dissect these performance gains through the lens of data memorization in overparameterized models.
no code implementations • 4 Jul 2023 • Sarah Sachs, Tim van Erven, Liam Hodgkinson, Rajiv Khanna, Umut Simsekli
Algorithm- and data-dependent generalization bounds are required to explain the generalization behavior of modern machine learning algorithms.
no code implementations • 24 Mar 2023 • Gregory Dexter, Rajiv Khanna, Jawad Raheel, Petros Drineas
We present novel bounds for coreset construction, feature selection, and dimensionality reduction for logistic regression.
no code implementations • 19 Feb 2023 • Kayhan Behdin, Qingquan Song, Aman Gupta, Sathiya Keerthi, Ayan Acharya, Borja Ocejo, Gregory Dexter, Rajiv Khanna, David Durfee, Rahul Mazumder
Modern deep learning models are over-parameterized, where different optima can result in widely varying generalization performance.
no code implementations • 28 Feb 2022 • Francesco Quinzan, Rajiv Khanna, Moshik Hershcovitch, Sarel Cohen, Daniel G. Waddington, Tobias Friedrich, Michael W. Mahoney
We study the fundamental problem of selecting optimal features for model construction.
no code implementations • 2 Aug 2021 • Liam Hodgkinson, Umut Şimşekli, Rajiv Khanna, Michael W. Mahoney
Despite the ubiquitous use of stochastic optimization algorithms in machine learning, the precise impact of these algorithms and their dynamics on generalization performance in realistic non-convex settings is still poorly understood.
no code implementations • 16 May 2021 • Vipul Gupta, Avishek Ghosh, Michal Derezinski, Rajiv Khanna, Kannan Ramchandran, Michael Mahoney
To enhance practicability, we devise an adaptive scheme to choose L, and we show that this reduces the number of local iterations in worker machines between two model synchronizations as the training proceeds, successively refining the model quality at the master.
no code implementations • NeurIPS 2020 • Michal Derezinski, Rajiv Khanna, Michael W. Mahoney
The Column Subset Selection Problem (CSSP) and the Nystrom method are among the leading tools for constructing small low-rank approximations of large datasets in machine learning and scientific computing.
1 code implementation • ICLR 2021 • Francisco Utrera, Evan Kravitz, N. Benjamin Erichson, Rajiv Khanna, Michael W. Mahoney
Transfer learning has emerged as a powerful methodology for adapting pre-trained deep neural networks on image recognition tasks to new domains.
1 code implementation • NeurIPS 2020 • Yaoqing Yang, Rajiv Khanna, Yaodong Yu, Amir Gholami, Kurt Keutzer, Joseph E. Gonzalez, Kannan Ramchandran, Michael W. Mahoney
Using these observations, we show that noise-augmentation on mixup training further increases boundary thickness, thereby combating vulnerability to various forms of adversarial attacks and OOD transforms.
1 code implementation • 1 Jul 2020 • Jacky Y. Zhang, Rajiv Khanna, Anastasios Kyrillidis, Oluwasanmi Koyejo
Bayesian coresets have emerged as a promising approach for implementing scalable Bayesian inference.
no code implementations • 21 Feb 2020 • Michał Dereziński, Rajiv Khanna, Michael W. Mahoney
The Column Subset Selection Problem (CSSP) and the Nystr\"om method are among the leading tools for constructing small low-rank approximations of large datasets in machine learning and scientific computing.
no code implementations • NeurIPS 2019 • Jacky Y. Zhang, Rajiv Khanna, Anastasios Kyrillidis, Oluwasanmi Koyejo
Iterative hard thresholding (IHT) is a projected gradient descent algorithm, known to achieve state of the art performance for a wide range of structured estimation problems, such as sparse inference.
no code implementations • 19 Jul 2019 • Rajiv Khanna, Liam Hodgkinson, Michael W. Mahoney
The rate of convergence of weighted kernel herding (WKH) and sequential Bayesian quadrature (SBQ), two kernel-based sampling algorithms for estimating integrals with respect to some target probability measure, is investigated.
no code implementations • 23 Oct 2018 • Rajiv Khanna, Been Kim, Joydeep Ghosh, Oluwasanmi Koyejo
Research in both machine learning and psychology suggests that salient examples can help humans to interpret learning models.
1 code implementation • NeurIPS 2018 • Francesco Locatello, Gideon Dresdner, Rajiv Khanna, Isabel Valera, Gunnar Rätsch
Finally, we present a stopping criterion drawn from the duality gap in the classic FW analyses and exhaustive experiments to illustrate the usefulness of our theoretical and algorithmic contributions.
no code implementations • 26 Dec 2017 • Rajiv Khanna, Anastasios Kyrillidis
We study --both in theory and practice-- the use of momentum motions in classic iterative hard thresholding (IHT) methods.
no code implementations • 5 Aug 2017 • Francesco Locatello, Rajiv Khanna, Joydeep Ghosh, Gunnar Rätsch
Variational inference is a popular technique to approximate a possibly intractable Bayesian posterior with a more tractable one.
no code implementations • ICML 2017 • Rajiv Khanna, Ethan Elenberg, Alexandros G. Dimakis, Sahand Negahban
We provide new approximation guarantees for greedy low rank matrix estimation under standard assumptions of restricted strong convexity and smoothness.
no code implementations • 8 Mar 2017 • Rajiv Khanna, Ethan Elenberg, Alexandros G. Dimakis, Sahand Negahban, Joydeep Ghosh
Furthermore, we show that a bounded submodularity ratio can be used to provide data dependent bounds that can sometimes be tighter also for submodular functions.
no code implementations • 21 Feb 2017 • Francesco Locatello, Rajiv Khanna, Michael Tschannen, Martin Jaggi
Two of the most fundamental prototypes of greedy optimization are the matching pursuit and Frank-Wolfe algorithms.
no code implementations • 2 Dec 2016 • Ethan R. Elenberg, Rajiv Khanna, Alexandros G. Dimakis, Sahand Negahban
Our results extend the work of Das and Kempe (2011) from the setting of linear regression to arbitrary objective functions.
no code implementations • NeurIPS 2016 • Been Kim, Rajiv Khanna, Oluwasanmi O. Koyejo
Example-based explanations are widely used in the effort to improve the interpretability of highly complex distributions.
no code implementations • 12 Jul 2016 • Rajiv Khanna, Joydeep Ghosh, Russell Poldrack, Oluwasanmi Koyejo
Approximate inference via information projection has been recently introduced as a general-purpose approach for efficient probabilistic inference given sparse variables.
no code implementations • 12 Feb 2016 • Rajiv Khanna, Michael Tschannen, Martin Jaggi
Efficiently representing real world data in a succinct and parsimonious manner is of central importance in many fields.
no code implementations • 6 Nov 2015 • S. Sathiya Keerthi, Tobias Schnabel, Rajiv Khanna
In a recent paper, Levy and Goldberg pointed out an interesting connection between prediction-based word embedding models and count models based on pointwise mutual information.
no code implementations • NeurIPS 2014 • Oluwasanmi O. Koyejo, Rajiv Khanna, Joydeep Ghosh, Russell Poldrack
In cases where this projection is intractable, we propose a family of parameterized approximations indexed by subsets of the domain.