no code implementations • 12 Feb 2024 • Yuval Filmus, Steve Hanneke, Idan Mehalel, Shay Moran
We demonstrate that the optimal mistake bound under bandit feedback is at most $O(k)$ times higher than the optimal mistake bound in the full information case, where $k$ represents the number of labels.
no code implementations • 8 Aug 2023 • Ananth Raman, Vinod Raman, Unique Subedi, Idan Mehalel, Ambuj Tewari
We study online multiclass classification under bandit feedback.
no code implementations • 27 Feb 2023 • Yuval Filmus, Steve Hanneke, Idan Mehalel, Shay Moran
We prove an analogous result for randomized learners: we show that the optimal expected mistake bound in learning a class $\mathcal{H}$ equals its randomized Littlestone dimension, which is the largest $d$ for which there exists a tree shattered by $\mathcal{H}$ whose average depth is $2d$.
no code implementations • 6 Oct 2022 • Steve Hanneke, Amin Karbasi, Mohammad Mahmoody, Idan Mehalel, Shay Moran
In this work we aim to characterize the smallest achievable error $\epsilon=\epsilon(\eta)$ by the learner in the presence of such an adversary in both realizable and agnostic settings.
no code implementations • 9 Jun 2022 • Yuval Filmus, Idan Mehalel, Shay Moran
Given a learning task where the data is distributed among several parties, communication is one of the fundamental resources which the parties would like to minimize.