Search Results for author: Vivak Patel

Found 6 papers, 0 papers with code

Global Convergence and Stability of Stochastic Gradient Descent

no code implementations4 Oct 2021 Vivak Patel, Shushu Zhang, Bowen Tian

Under a slightly more restrictive assumption on the joint behavior of the non-convexity and noise model that generalizes current assumptions in the literature, we show that the objective function cannot diverge, even if the iterates diverge.

Stochastic Optimization

Stochastic Approximation for High-frequency Observations in Data Assimilation

no code implementations5 Nov 2020 Shushu Zhang, Vivak Patel

With the increasing penetration of high-frequency sensors across a number of biological and physical systems, the abundance of the resulting observations offers opportunities for higher statistical accuracy of down-stream estimates, but their frequency results in a plethora of computational problems in data assimilation tasks.

Vocal Bursts Intensity Prediction

Stopping Criteria for, and Strong Convergence of, Stochastic Gradient Descent on Bottou-Curtis-Nocedal Functions

no code implementations1 Apr 2020 Vivak Patel

As a result of our work, our rigorously developed stopping criteria can be used to develop new adaptive step size schemes or bolster other downstream analyses for nonconvex functions.

Open-Ended Question Answering

The Impact of Local Geometry and Batch Size on Stochastic Gradient Descent for Nonconvex Problems

no code implementations14 Sep 2017 Vivak Patel

In several experimental reports on nonconvex optimization problems in machine learning, stochastic gradient descent (SGD) was observed to prefer minimizers with flat basins in comparison to more deterministic methods, yet there is very little rigorous understanding of this phenomenon.

On SGD's Failure in Practice: Characterizing and Overcoming Stalling

no code implementations1 Feb 2017 Vivak Patel

Stochastic Gradient Descent (SGD) is widely used in machine learning problems to efficiently perform empirical risk minimization, yet, in practice, SGD is known to stall before reaching the actual minimizer of the empirical risk.

Unity

Kalman-based Stochastic Gradient Method with Stop Condition and Insensitivity to Conditioning

no code implementations3 Dec 2015 Vivak Patel

Modern proximal and stochastic gradient descent (SGD) methods are believed to efficiently minimize large composite objective functions, but such methods have two algorithmic challenges: (1) a lack of fast or justified stop conditions, and (2) sensitivity to the objective function's conditioning.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.