Search Results for author: Vashist Avadhanula

Found 10 papers, 0 papers with code

Fully Dynamic Online Selection through Online Contention Resolution Schemes

no code implementations8 Jan 2023 Vashist Avadhanula, Andrea Celli, Riccardo Colini-Baldeschi, Stefano Leonardi, Matteo Russo

A successful approach to online selection problems in the adversarial setting is given by the notion of Online Contention Resolution Scheme (OCRS), that uses a priori information to formulate a linear relaxation of the underlying optimization problem, whose optimal fractional solution is rounded online for any adversarial order of the input sequence.

Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations

no code implementations13 Dec 2021 Evrard Garcelon, Vashist Avadhanula, Alessandro Lazaric, Matteo Pirotta

We consider a multi-armed bandit setting where, at the beginning of each round, the learner receives noisy independent, and possibly biased, \emph{evaluations} of the true reward of each arm and it selects $K$ arms with the objective of accumulating as much reward as possible over $T$ rounds.

QUEST: Queue Simulation for Content Moderation at Scale

no code implementations31 Mar 2021 Rahul Makhijani, Parikshit Shah, Vashist Avadhanula, Caner Gocmen, Nicolás E. Stier-Moses, Julián Mestre

Moderating content in social media platforms is a formidable challenge due to the unprecedented scale of such systems, which typically handle billions of posts per day.

BIG-bench Machine Learning

Stochastic Bandits for Multi-platform Budget Optimization in Online Advertising

no code implementations16 Mar 2021 Vashist Avadhanula, Riccardo Colini-Baldeschi, Stefano Leonardi, Karthik Abinav Sankararaman, Okke Schrijvers

We modify the algorithm proposed in Badanidiyuru \emph{et al.,} to extend it to the case of multiple platforms to obtain an algorithm for both the discrete and continuous bid-spaces.

A Tractable Online Learning Algorithm for the Multinomial Logit Contextual Bandit

no code implementations28 Nov 2020 Priyank Agrawal, Theja Tulabandhula, Vashist Avadhanula

In this paper, we propose an optimistic algorithm and show that the regret is bounded by $O(\sqrt{dT} + \kappa)$, significantly improving the performance over existing methods.

Decision Making Multi-Armed Bandits

Multi-armed Bandits with Cost Subsidy

no code implementations3 Nov 2020 Deeksha Sinha, Karthik Abinav Sankararama, Abbas Kazerouni, Vashist Avadhanula

We then establish a fundamental lower bound on the performance of any online learning algorithm for this problem, highlighting the hardness of our problem in comparison to the classical MAB problem.

Multi-Armed Bandits Thompson Sampling

MNL-Bandit: A Dynamic Learning Approach to Assortment Selection

no code implementations13 Jun 2017 Shipra Agrawal, Vashist Avadhanula, Vineet Goyal, Assaf Zeevi

The retailer observes this choice and the objective is to dynamically learn the model parameters, while optimizing cumulative revenues over a selling horizon of length $T$.

Thompson Sampling for the MNL-Bandit

no code implementations3 Jun 2017 Shipra Agrawal, Vashist Avadhanula, Vineet Goyal, Assaf Zeevi

We consider a sequential subset selection problem under parameter uncertainty, where at each time step, the decision maker selects a subset of cardinality $K$ from $N$ possible items (arms), and observes a (bandit) feedback in the form of the index of one of the items in said subset, or none.

Thompson Sampling

Cannot find the paper you are looking for? You can Submit a new open access paper.