Search Results for author: Krishnakumar Balasubramanian

Found 51 papers, 2 papers with code

On the Optimality of Kernel-Embedding Based Goodness-of-Fit Tests

no code implementations24 Sep 2017 Krishnakumar Balasubramanian, Tong Li, Ming Yuan

The reproducing kernel Hilbert space (RKHS) embedding of distributions offers a general and flexible framework for testing problems in arbitrary domains and has attracted considerable amount of attention in recent years.

On Stein's Identity and Near-Optimal Estimation in High-dimensional Index Models

no code implementations26 Sep 2017 Zhuoran Yang, Krishnakumar Balasubramanian, Han Liu

We consider estimating the parametric components of semi-parametric multiple index models in a high-dimensional and non-Gaussian setting.

Tensor Methods for Additive Index Models under Discordance and Heterogeneity

no code implementations17 Jul 2018 Krishnakumar Balasubramanian, Jianqing Fan, Zhuoran Yang

Motivated by the sampling problems and heterogeneity issues common in high- dimensional big datasets, we consider a class of discordant additive index models.

Zeroth-order Nonconvex Stochastic Optimization: Handling Constraints, High-Dimensionality and Saddle-Points

no code implementations NeurIPS 2018 Krishnakumar Balasubramanian, Saeed Ghadimi

In this paper, we propose and analyze zeroth-order stochastic approximation algorithms for nonconvex and convex optimization, with a focus on addressing constrained optimization, high-dimensional setting and saddle-point avoiding.

Stochastic Optimization Vocal Bursts Intensity Prediction

Stochastic Zeroth-order Discretizations of Langevin Diffusions for Bayesian Inference

no code implementations4 Feb 2019 Abhishek Roy, Lingqing Shen, Krishnakumar Balasubramanian, Saeed Ghadimi

Our theoretical contributions extend the practical applicability of sampling algorithms to the noisy black-box and high-dimensional settings.

Bayesian Inference Stochastic Optimization +1

Normal Approximation for Stochastic Gradient Descent via Non-Asymptotic Rates of Martingale CLT

no code implementations3 Apr 2019 Andreas Anastasiou, Krishnakumar Balasubramanian, Murat A. Erdogdu

A crucial intermediate step is proving a non-asymptotic martingale central limit theorem (CLT), i. e., establishing the rates of convergence of a multivariate martingale difference sequence to a normal random vector, which might be of independent interest.

valid

Multi-Point Bandit Algorithms for Nonstationary Online Nonconvex Optimization

no code implementations31 Jul 2019 Abhishek Roy, Krishnakumar Balasubramanian, Saeed Ghadimi, Prasant Mohapatra

In this paper, motivated by online reinforcement learning problems, we propose and analyze bandit algorithms for both general and structured nonconvex problems with nonstationary (or dynamic) regret as the performance measure, in both stochastic and non-stochastic settings.

Stochastic Zeroth-order Riemannian Derivative Estimation and Optimization

no code implementations25 Mar 2020 Jiaxiang Li, Krishnakumar Balasubramanian, Shiqian Ma

We consider stochastic zeroth-order optimization over Riemannian submanifolds embedded in Euclidean space, where the task is to solve Riemannian optimization problem with only noisy objective function evaluations.

Riemannian optimization

An Analysis of Constant Step Size SGD in the Non-convex Regime: Asymptotic Normality and Bias

no code implementations NeurIPS 2021 Lu Yu, Krishnakumar Balasubramanian, Stanislav Volgushev, Murat A. Erdogdu

Structured non-convex learning problems, for which critical points have favorable statistical properties, arise frequently in statistical machine learning.

Improved Complexities for Stochastic Conditional Gradient Methods under Interpolation-like Conditions

no code implementations15 Jun 2020 Tesi Xiao, Krishnakumar Balasubramanian, Saeed Ghadimi

We analyze stochastic conditional gradient methods for constrained optimization problems arising in over-parametrized machine learning.

BIG-bench Machine Learning

Fractal Gaussian Networks: A sparse random graph model based on Gaussian Multiplicative Chaos

no code implementations ICML 2020 Subhroshekhar Ghosh, Krishnakumar Balasubramanian, Xiaochuan Yang

We propose a novel stochastic network model, called Fractal Gaussian Network (FGN), that embodies well-defined and analytically tractable fractal structures.

Stochastic Block Model

Stochastic Multi-level Composition Optimization Algorithms with Level-Independent Convergence Rates

no code implementations24 Aug 2020 Krishnakumar Balasubramanian, Saeed Ghadimi, Anthony Nguyen

We show that the first algorithm, which is a generalization of \cite{GhaRuswan20} to the $T$ level case, can achieve a sample complexity of $\mathcal{O}(1/\epsilon^6)$ by using mini-batches of samples in each iteration.

Escaping Saddle-Points Faster under Interpolation-like Conditions

no code implementations28 Sep 2020 Abhishek Roy, Krishnakumar Balasubramanian, Saeed Ghadimi, Prasant Mohapatra

We next analyze Stochastic Cubic-Regularized Newton (SCRN) algorithm under interpolation-like conditions, and show that the oracle complexity to reach an $\epsilon$-local-minimizer under interpolation-like conditions, is $\tilde{\mathcal{O}}(1/\epsilon^{2. 5})$.

Stochastic Optimization

On the Ergodicity, Bias and Asymptotic Normality of Randomized Midpoint Sampling Method

no code implementations NeurIPS 2020 Ye He, Krishnakumar Balasubramanian, Murat A. Erdogdu

The randomized midpoint method, proposed by [SL19], has emerged as an optimal discretization procedure for simulating the continuous time Langevin diffusions.

Numerical Integration

Escaping Saddle-Point Faster under Interpolation-like Conditions

no code implementations NeurIPS 2020 Abhishek Roy, Krishnakumar Balasubramanian, Saeed Ghadimi, Prasant Mohapatra

We next analyze Stochastic Cubic-Regularized Newton (SCRN) algorithm under interpolation-like conditions, and show that the oracle complexity to reach an $\epsilon$-local-minimizer under interpolation-like conditions, is $O(1/\epsilon^{2. 5})$.

Stochastic Optimization

Statistical Inference for Polyak-Ruppert Averaged Zeroth-order Stochastic Gradient Algorithm

no code implementations10 Feb 2021 Yanhao Jin, Tesi Xiao, Krishnakumar Balasubramanian

Statistical machine learning models trained with stochastic gradient algorithms are increasingly being deployed in critical scientific applications.

BIG-bench Machine Learning valid

Nonparametric Modeling of Higher-Order Interactions via Hypergraphons

no code implementations18 May 2021 Krishnakumar Balasubramanian

We study statistical and algorithmic aspects of using hypergraphons, that are limits of large hypergraphs, for modeling higher-order interactions.

On Empirical Risk Minimization with Dependent and Heavy-Tailed Data

no code implementations NeurIPS 2021 Abhishek Roy, Krishnakumar Balasubramanian, Murat A. Erdogdu

In this work, we establish risk bounds for the Empirical Risk Minimization (ERM) with both dependent and heavy-tailed data-generating processes.

Learning Theory

Topologically penalized regression on manifolds

no code implementations26 Oct 2021 Olympio Hacquard, Krishnakumar Balasubramanian, Gilles Blanchard, Clément Levrard, Wolfgang Polonik

We study a regression problem on a compact manifold M. In order to take advantage of the underlying geometry and topology of the data, the regression task is performed on the basis of the first several eigenfunctions of the Laplace-Beltrami operator of the manifold, that are regularized with topological penalties.

regression

Heavy-tailed Sampling via Transformed Unadjusted Langevin Algorithm

no code implementations20 Jan 2022 Ye He, Krishnakumar Balasubramanian, Murat A. Erdogdu

We analyze the oracle complexity of sampling from polynomially decaying heavy-tailed target densities based on running the Unadjusted Langevin Algorithm on certain transformed versions of the target density.

A Projection-free Algorithm for Constrained Stochastic Multi-level Composition Optimization

no code implementations9 Feb 2022 Tesi Xiao, Krishnakumar Balasubramanian, Saeed Ghadimi

We propose a projection-free conditional gradient-type algorithm for smooth stochastic multi-level composition optimization, where the objective function is a nested composition of $T$ functions and the constraint set is a closed convex set.

Towards a Theory of Non-Log-Concave Sampling: First-Order Stationarity Guarantees for Langevin Monte Carlo

no code implementations10 Feb 2022 Krishnakumar Balasubramanian, Sinho Chewi, Murat A. Erdogdu, Adil Salim, Matthew Zhang

For the task of sampling from a density $\pi \propto \exp(-V)$ on $\mathbb{R}^d$, where $V$ is possibly non-convex but $L$-gradient Lipschitz, we prove that averaged Langevin Monte Carlo outputs a sample with $\varepsilon$-relative Fisher information after $O( L^2 d^2/\varepsilon^2)$ iterations.

A Flexible Approach for Normal Approximation of Geometric and Topological Statistics

no code implementations19 Oct 2022 Zhaoyang Shi, Krishnakumar Balasubramanian, Wolfgang Polonik

We derive normal approximation results for a class of stabilizing functionals of binomial or Poisson point process, that are not necessarily expressible as sums of certain score functions.

Decentralized Stochastic Bilevel Optimization with Improved per-Iteration Complexity

no code implementations23 Oct 2022 Xuxing Chen, Minhui Huang, Shiqian Ma, Krishnakumar Balasubramanian

Bilevel optimization recently has received tremendous attention due to its great success in solving important machine learning problems like meta learning, reinforcement learning, and hyperparameter optimization.

Bilevel Optimization Hyperparameter Optimization +2

Regularized Stein Variational Gradient Flow

no code implementations15 Nov 2022 Ye He, Krishnakumar Balasubramanian, Bharath K. Sriperumbudur, Jianfeng Lu

In this work, we propose the Regularized Stein Variational Gradient Flow which interpolates between the Stein Variational Gradient Flow and the Wasserstein Gradient Flow.

Improved Discretization Analysis for Underdamped Langevin Monte Carlo

no code implementations16 Feb 2023 Matthew Zhang, Sinho Chewi, Mufan Bill Li, Krishnakumar Balasubramanian, Murat A. Erdogdu

As a byproduct, we also obtain the first KL divergence guarantees for ULMC without Hessian smoothness under strong log-concavity, which is based on a new result on the log-Sobolev constant along the underdamped Langevin diffusion.

A One-Sample Decentralized Proximal Algorithm for Non-Convex Stochastic Composite Optimization

1 code implementation20 Feb 2023 Tesi Xiao, Xuxing Chen, Krishnakumar Balasubramanian, Saeed Ghadimi

We focus on decentralized stochastic non-convex optimization, where $n$ agents work together to optimize a composite objective function which is a sum of a smooth term and a non-smooth convex term.

Mean-Square Analysis of Discretized Itô Diffusions for Heavy-tailed Sampling

no code implementations1 Mar 2023 Ye He, Tyler Farghly, Krishnakumar Balasubramanian, Murat A. Erdogdu

We analyze the complexity of sampling from a class of heavy-tailed distributions by discretizing a natural class of It\^o diffusions associated with weighted Poincar\'e inequalities.

Towards a Complete Analysis of Langevin Monte Carlo: Beyond Poincaré Inequality

no code implementations7 Mar 2023 Alireza Mousavi-Hosseini, Tyler Farghly, Ye He, Krishnakumar Balasubramanian, Murat A. Erdogdu

We do so by establishing upper and lower bounds for Langevin diffusions and LMC under weak Poincar\'e inequalities that are satisfied by a large class of densities including polynomially-decaying heavy-tailed densities (i. e., Cauchy-type).

High-dimensional scaling limits and fluctuations of online least-squares SGD with smooth covariance

no code implementations3 Apr 2023 Krishnakumar Balasubramanian, Promit Ghosal, Ye He

We derive high-dimensional scaling limits and fluctuations for the online least-squares Stochastic Gradient Descent (SGD) algorithm by taking the properties of the data generating model explicitly into consideration.

Forward-backward Gaussian variational inference via JKO in the Bures-Wasserstein Space

no code implementations10 Apr 2023 Michael Diao, Krishnakumar Balasubramanian, Sinho Chewi, Adil Salim

Of key interest in statistics and machine learning is Gaussian VI, which approximates $\pi$ by minimizing the Kullback-Leibler (KL) divergence to $\pi$ over the space of Gaussians.

Variational Inference

Optimal Algorithms for Stochastic Bilevel Optimization under Relaxed Smoothness Conditions

no code implementations21 Jun 2023 Xuxing Chen, Tesi Xiao, Krishnakumar Balasubramanian

In this paper, we introduce a novel fully single-loop and Hessian-inversion-free algorithmic framework for stochastic bilevel optimization and present a tighter analysis under standard smoothness assumptions (first-order Lipschitzness of the UL function and second-order Lipschitzness of the LL function).

Bilevel Optimization

Gaussian random field approximation via Stein's method with applications to wide random neural networks

no code implementations28 Jun 2023 Krishnakumar Balasubramanian, Larry Goldstein, Nathan Ross, Adil Salim

Specializing our general result, we obtain the first bounds on the Gaussian random field approximation of wide random neural networks of any depth and Lipschitz activation functions at the random field level.

Stochastic Nested Compositional Bi-level Optimization for Robust Feature Learning

no code implementations11 Jul 2023 Xuxing Chen, Krishnakumar Balasubramanian, Saeed Ghadimi

We develop and analyze stochastic approximation algorithms for solving nested compositional bi-level optimization problems.

Online covariance estimation for stochastic gradient descent under Markovian sampling

no code implementations3 Aug 2023 Abhishek Roy, Krishnakumar Balasubramanian

We investigate the online overlapping batch-means covariance estimator for Stochastic Gradient Descent (SGD) under Markovian sampling.

regression

Zeroth-order Riemannian Averaging Stochastic Approximation Algorithms

no code implementations25 Sep 2023 Jiaxiang Li, Krishnakumar Balasubramanian, Shiqian Ma

We present Zeroth-order Riemannian Averaging Stochastic Approximation (\texttt{Zo-RASA}) algorithms for stochastic optimization on Riemannian manifolds.

Stochastic Optimization

From Stability to Chaos: Analyzing Gradient Descent Dynamics in Quadratic Regression

no code implementations2 Oct 2023 Xuxing Chen, Krishnakumar Balasubramanian, Promit Ghosal, Bhavya Agrawalla

We conduct a comprehensive investigation into the dynamics of gradient descent using large-order constant step-sizes in the context of quadratic regression models.

regression Retrieval

Adaptive and non-adaptive minimax rates for weighted Laplacian-eigenmap based nonparametric regression

no code implementations31 Oct 2023 Zhaoyang Shi, Krishnakumar Balasubramanian, Wolfgang Polonik

We show both adaptive and non-adaptive minimax rates of convergence for a family of weighted Laplacian-Eigenmap based nonparametric regression methods, when the true regression function belongs to a Sobolev space and the sampling density is bounded from above and below.

regression

Nonsmooth Nonparametric Regression via Fractional Laplacian Eigenmaps

no code implementations22 Feb 2024 Zhaoyang Shi, Krishnakumar Balasubramanian, Wolfgang Polonik

More specifically, our approach is using the fractional Laplacian and is designed to handle the case when the true regression function lies in an $L_2$-fractional Sobolev space with order $s\in (0, 1)$.

regression

Multivariate Gaussian Approximation for Random Forest via Region-based Stabilization

no code implementations15 Mar 2024 Zhaoyang Shi, Chinmoy Bhattacharjee, Krishnakumar Balasubramanian, Wolfgang Polonik

We derive Gaussian approximation bounds for random forest predictions based on a set of training points given by a Poisson process, under fairly mild regularity assumptions on the data generating process.

Meta-Learning with Generalized Ridge Regression: High-dimensional Asymptotics, Optimality and Hyper-covariance Estimation

1 code implementation27 Mar 2024 Yanhao Jin, Krishnakumar Balasubramanian, Debashis Paul

Finally, we propose and analyze an estimator of the inverse covariance matrix of random regression coefficients based on data from the training tasks.

Meta-Learning regression +1

Minimax Optimal Goodness-of-Fit Testing with Kernel Stein Discrepancy

no code implementations12 Apr 2024 Omar Hagrass, Bharath Sriperumbudur, Krishnakumar Balasubramanian

We explore the minimax optimality of goodness-of-fit tests on general domains using the kernelized Stein discrepancy (KSD).

Computational Efficiency

Cannot find the paper you are looking for? You can Submit a new open access paper.