1 code implementation • 5 Jun 2023 • Boxin Zhao, Boxiang Lyu, Raul Castro Fernandez, Mladen Kolar
Data markets help with identifying valuable training data: model consumers pay to train a model, the market uses that budget to identify data and train the model (the budget allocation problem), and finally the market compensates data providers according to their data contribution (revenue allocation problem).
no code implementations • 1 Jun 2023 • Boxiang Lyu, Zhe Feng, Zachary Robertson, Sanmi Koyejo
We study the design of loss functions for click-through rates (CTR) to optimize (social) welfare in advertising auctions.
no code implementations • 19 Oct 2022 • Rui Ai, Boxiang Lyu, Zhaoran Wang, Zhuoran Yang, Michael I. Jordan
First, from the seller's perspective, we need to efficiently explore the environment in the presence of potentially nontruthful bidders who aim to manipulates seller's policy.
no code implementations • 31 May 2022 • Pedro Cisneros-Velarde, Boxiang Lyu, Sanmi Koyejo, Mladen Kolar
Although parallelism has been extensively used in reinforcement learning (RL), the quantitative effects of parallel exploration are not well understood theoretically.
no code implementations • 5 May 2022 • Boxiang Lyu, Zhaoran Wang, Mladen Kolar, Zhuoran Yang
In the setting where the function approximation is employed to handle large state spaces, with only mild assumptions on the expressiveness of the function class, we are able to design a dynamic mechanism using offline reinforcement learning algorithms.
1 code implementation • 28 Apr 2022 • Boxiang Lyu, Filip Hanzely, Mladen Kolar
We consider the problem of personalized federated learning when there are known cluster structures within users.
no code implementations • 25 Feb 2022 • Shuang Qiu, Boxiang Lyu, Qinglin Meng, Zhaoran Wang, Zhuoran Yang, Michael I. Jordan
Dynamic mechanism design studies how mechanism designers should allocate resources among agents in a time-varying environment.
no code implementations • 31 Jan 2022 • Boxin Zhao, Boxiang Lyu, Mladen Kolar
Stochastic gradient-based optimization methods, such as L-SVRG and its accelerated variant L-Katyusha (Kovalev et al., 2020), are widely used to train machine learning models. The theoretical and empirical performance of L-SVRG and L-Katyusha can be improved by sampling observations from a non-uniform distribution (Qian et al., 2021).