Search Results for author: Aristide Tossou

Found 7 papers, 0 papers with code

Near-optimal Bayesian Solution For Unknown Discrete Markov Decision Process

no code implementations20 Jun 2019 Aristide Tossou, Christos Dimitrakakis, Debabrota Basu

We derive the first polynomial time Bayesian algorithm, BUCRL{} that achieves up to logarithm factors, a regret (i. e the difference between the accumulated rewards of the optimal policy and our algorithm) of the optimal order $\tilde{\mathcal{O}}(\sqrt{DSAT})$.

Near-Optimal Online Egalitarian learning in General Sum Repeated Matrix Games

no code implementations4 Jun 2019 Aristide Tossou, Christos Dimitrakakis, Jaroslaw Rzepecki, Katja Hofmann

We study two-player general sum repeated finite games where the rewards of each player are generated from an unknown distribution.

Differential Privacy for Multi-armed Bandits: What Is It and What Is Its Cost?

no code implementations29 May 2019 Debabrota Basu, Christos Dimitrakakis, Aristide Tossou

We derive and contrast lower bounds on the regret of bandit algorithms satisfying these definitions.

Multi-Armed Bandits

Learning to Match

no code implementations30 Jul 2017 Philip Ekman, Sebastian Bellevik, Christos Dimitrakakis, Aristide Tossou

One specific such problem involves matching a set of workers to a set of tasks.

Algorithms for Differentially Private Multi-Armed Bandits

no code implementations27 Nov 2015 Aristide Tossou, Christos Dimitrakakis

This is a significant improvement over previous results, which only achieve poly-log regret $O(\epsilon^{-2} \log^{2} T)$, because of our use of a novel interval-based mechanism.

Multi-Armed Bandits

Probabilistic inverse reinforcement learning in unknown environments

no code implementations9 Aug 2014 Aristide Tossou, Christos Dimitrakakis

To do so, we extend previous probabilistic approaches for inverse reinforcement learning in known MDPs to the case of unknown dynamics or opponents.

Bayesian Inference reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.