1 code implementation • NeurIPS 2023 • Parnian Kassraie, Nicolas Emmenegger, Andreas Krause, Aldo Pacchiano
This allows us to develop ALEXP, which has an exponentially improved ($\log M$) dependence on $M$ for its regret.
1 code implementation • 2 Mar 2023 • Jonas Rothfuss, Bhavya Sukhija, Tobias Birchler, Parnian Kassraie, Andreas Krause
We study the problem of conservative off-policy evaluation (COPE) where given an offline dataset of environment interactions, collected by other agents, we seek to obtain a (tight) lower bound on a policy's performance.
no code implementations • 2 Nov 2022 • Songyan Hou, Parnian Kassraie, Anastasis Kratsios, Andreas Krause, Jonas Rothfuss
Existing generalization bounds fail to explain crucial factors that drive the generalization of modern neural networks.
no code implementations • 27 Oct 2022 • Felix Schur, Parnian Kassraie, Jonas Rothfuss, Andreas Krause
Our algorithm can be paired with any kernelized or linear bandit algorithm and guarantees oracle optimal performance, meaning that as more tasks are solved, the regret of LIBO on each task converges to the regret of the bandit algorithm with oracle knowledge of the true kernel.
no code implementations • 13 Jul 2022 • Parnian Kassraie, Andreas Krause, Ilija Bogunovic
By establishing a novel connection between such kernels and the graph neural tangent kernel (GNTK), we introduce the first GNN confidence bound and use it to design a phased-elimination algorithm with sublinear regret.
no code implementations • 1 Feb 2022 • Parnian Kassraie, Jonas Rothfuss, Andreas Krause
We demonstrate our approach on the kernelized bandit problem (a. k. a.~Bayesian optimization), where we establish regret bounds competitive with those given the true kernel.
1 code implementation • 7 Jul 2021 • Parnian Kassraie, Andreas Krause
Contextual bandits are a rich model for sequential decision making given side information, with important applications, e. g., in recommender systems.