no code implementations • ICML 2020 • Pooria Joulani, Anant Raj, András György, Csaba Szepesvari
In this paper, we show that there is a simpler approach to obtaining accelerated rates: applying generic, well-known optimistic online learning algorithms and using the online average of their predictions to query the (deterministic or stochastic) first-order optimization oracle at each time step.
no code implementations • 12 Oct 2020 • András György, Pooria Joulani
We consider the adversarial multi-armed bandit problem under delayed feedback.
1 code implementation • 8 Feb 2020 • Botao Hao, Nevena Lazic, Yasin Abbasi-Yadkori, Pooria Joulani, Csaba Szepesvari
This is an improvement over the best existing bound of $\tilde{O}(T^{3/4})$ for the average-reward case with function approximation.
no code implementations • NeurIPS 2019 • Pooria Joulani, András György, Csaba Szepesvari
ASYNCADA is, to our knowledge, the first asynchronous stochastic optimization algorithm with finite-time data-dependent convergence guarantees for generic convex constraints.
no code implementations • 8 Sep 2017 • Pooria Joulani, András György, Csaba Szepesvári
Recently, much work has been done on extending the scope of online learning and incremental stochastic optimization algorithms.
no code implementations • 30 Jun 2015 • Pooria Joulani, András György, Csaba Szepesvári
Cross-validation (CV) is one of the main tools for performance estimation and parameter tuning in machine learning.
no code implementations • 4 Jun 2013 • Pooria Joulani, András György, Csaba Szepesvári
Online learning with delayed feedback has received increasing attention recently due to its several applications in distributed, web-based learning problems.