Search Results for author: Nicholas J. A. Harvey

Found 8 papers, 0 papers with code

Continuous Prediction with Experts' Advice

no code implementations1 Jun 2022 Victor Sanches Portella, Christopher Liaw, Nicholas J. A. Harvey

Finally, we design an anytime continuous-time algorithm with regret matching the optimal fixed-time rate when the gains are independent Brownian Motions; in many settings, this is the most difficult case.

Efficient and Optimal Fixed-Time Regret with Two Experts

no code implementations15 Mar 2022 Laura Greenstreet, Nicholas J. A. Harvey, Victor Sanches Portella

In instances with $T$ rounds and $n$ experts, the classical Multiplicative Weights Update method suffers at most $\sqrt{(T/2)\ln n}$ regret when $T$ is known beforehand.

Vocal Bursts Valence Prediction

Online mirror descent and dual averaging: keeping pace in the dynamic case

no code implementations ICML 2020 Huang Fang, Nicholas J. A. Harvey, Victor S. Portella, Michael P. Friedlander

Online mirror descent (OMD) and dual averaging (DA) -- two fundamental algorithms for online convex optimization -- are known to have very similar (and sometimes identical) performance guarantees when used with a fixed learning rate.

Optimal anytime regret with two experts

no code implementations20 Feb 2020 Nicholas J. A. Harvey, Christopher Liaw, Edwin Perkins, Sikander Randhawa

In the fixed-time setting, where the time horizon is known in advance, algorithms that achieve the optimal regret are known when there are two, three, or four experts or when the number of experts is large.

Vocal Bursts Valence Prediction

Simple and optimal high-probability bounds for strongly-convex stochastic gradient descent

no code implementations2 Sep 2019 Nicholas J. A. Harvey, Christopher Liaw, Sikander Randhawa

We consider a simple, non-uniform averaging strategy of Lacoste-Julien et al. (2011) and prove that it achieves the optimal $O(1/T)$ convergence rate with high probability.

Tight Analyses for Non-Smooth Stochastic Gradient Descent

no code implementations13 Dec 2018 Nicholas J. A. Harvey, Christopher Liaw, Yaniv Plan, Sikander Randhawa

We prove that after $T$ steps of stochastic gradient descent, the error of the final iterate is $O(\log(T)/T)$ with high probability.

The complexity of UNO

no code implementations15 Mar 2010 Erik D. Demaine, Martin L. Demaine, Nicholas J. A. Harvey, Ryuhei Uehara, Takeaki Uno, Yushi Uno

This paper investigates the popular card game UNO from the viewpoint of algorithmic combinatorial game theory.

Discrete Mathematics Computational Complexity G.2; F.1

Cannot find the paper you are looking for? You can Submit a new open access paper.