Search Results for author: Digvijay Boob

Found 12 papers, 2 papers with code

Complexity of Training ReLU Neural Networks

no code implementations ICLR 2019 Digvijay Boob, Santanu S. Dey, Guanghui Lan

In this paper, we explore some basic questions on complexity of training Neural networks with ReLU activation function.

First-order methods for Stochastic Variational Inequality problems with Function Constraints

no code implementations10 Apr 2023 Digvijay Boob, Qi Deng

Second, to obtain the optimal operator complexity for smooth deterministic problems, we present a novel single-loop Adaptive Lagrangian Extrapolation~(\texttt{AdLagEx}) method that can adaptively search for and explicitly bound the Lagrange multipliers.

Accelerated Primal-Dual Methods for Convex-Strongly-Concave Saddle Point Problems

no code implementations10 Sep 2022 Mohammad Khalafi, Digvijay Boob

We investigate a primal-dual (PD) method for the saddle point problem (SPP) that uses a linear approximation of the primal function instead of the standard proximal step, resulting in a linearized PD (LPD) method.

Optimal Algorithms for Differentially Private Stochastic Monotone Variational Inequalities and Saddle-Point Problems

no code implementations7 Apr 2021 Digvijay Boob, Cristóbal Guzmán

We show that a stochastic approximation variant of these algorithms attains risk bounds vanishing as a function of the dataset size, with respect to the strong gap function; and a sampling with replacement variant achieves optimal risk bounds with respect to a weak gap function.

A Feasible Level Proximal Point Method for Nonconvex Sparse Constrained Optimization

no code implementations NeurIPS 2020 Digvijay Boob, Qi Deng, Guanghui Lan, Yilin Wang

We also establish new convergence complexities to achieve an approximate KKT solution when the objective can be smooth/nonsmooth, deterministic/stochastic and convex/nonconvex with complexity that is on a par with gradient descent for unconstrained optimization problems in respective cases.

Differentially Private Synthetic Mixed-Type Data Generation For Unsupervised Learning

1 code implementation6 Dec 2019 Uthaipon Tantipongpipat, Chris Waites, Digvijay Boob, Amaresh Ankit Siva, Rachel Cummings

We introduce the DP-auto-GAN framework for synthetic data generation, which combines the low dimensional representation of autoencoders with the flexibility of Generative Adversarial Networks (GANs).

Synthetic Data Generation Vocal Bursts Type Prediction

Faster width-dependent algorithm for mixed packing and covering LPs

no code implementations NeurIPS 2019 Digvijay Boob, Saurabh Sawlani, Di Wang

As a special case of our result, we report a $1+\eps$ approximation algorithm for the densest subgraph problem which runs in time $O(md/ \eps)$, where $m$ is the number of edges in the graph and $d$ is the maximum graph degree.

Combinatorial Optimization

Differentially Private Mixed-Type Data Generation For Unsupervised Learning

1 code implementation25 Sep 2019 Uthaipon Tantipongpipat, Chris Waites, Digvijay Boob, Amaresh Siva, Rachel Cummings

In this work we introduce the DP-auto-GAN framework for synthetic data generation, which combines the low dimensional representation of autoencoders with the flexibility of GANs.

Synthetic Data Generation Vocal Bursts Type Prediction

Stochastic First-order Methods for Convex and Nonconvex Functional Constrained Optimization

no code implementations7 Aug 2019 Digvijay Boob, Qi Deng, Guanghui Lan

For large-scale and stochastic problems, we present a more practical proximal point method in which the approximate solutions of the subproblems are computed by the aforementioned ConEx method.

BIG-bench Machine Learning

Complexity of Training ReLU Neural Network

no code implementations27 Sep 2018 Digvijay Boob, Santanu S. Dey, Guanghui Lan

In this paper, we explore some basic questions on the complexity of training neural networks with ReLU activation function.

Theoretical properties of the global optimizer of two-layer Neural Network

no code implementations ICLR 2018 Digvijay Boob, Guanghui Lan

We essentially show that these non-singular hidden layer matrix satisfy a ``"good" property for these big class of activation functions.

Vocal Bursts Valence Prediction

Theoretical properties of the global optimizer of two layer neural network

no code implementations30 Oct 2017 Digvijay Boob, Guanghui Lan

We look at this problem in the setting where the number of parameters is greater than the number of sampled points.

Vocal Bursts Valence Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.