Search Results for author: Boxiang Lyu

Found 8 papers, 2 papers with code

Addressing Budget Allocation and Revenue Allocation in Data Market Environments Using an Adaptive Sampling Algorithm

1 code implementation5 Jun 2023 Boxin Zhao, Boxiang Lyu, Raul Castro Fernandez, Mladen Kolar

Data markets help with identifying valuable training data: model consumers pay to train a model, the market uses that budget to identify data and train the model (the budget allocation problem), and finally the market compensates data providers according to their data contribution (revenue allocation problem).

Fraud Detection

Pairwise Ranking Losses of Click-Through Rates Prediction for Welfare Maximization in Ad Auctions

no code implementations1 Jun 2023 Boxiang Lyu, Zhe Feng, Zachary Robertson, Sanmi Koyejo

We study the design of loss functions for click-through rates (CTR) to optimize (social) welfare in advertising auctions.

Learning-To-Rank

A Reinforcement Learning Approach in Multi-Phase Second-Price Auction Design

no code implementations19 Oct 2022 Rui Ai, Boxiang Lyu, Zhaoran Wang, Zhuoran Yang, Michael I. Jordan

First, from the seller's perspective, we need to efficiently explore the environment in the presence of potentially nontruthful bidders who aim to manipulates seller's policy.

reinforcement-learning Reinforcement Learning (RL)

One Policy is Enough: Parallel Exploration with a Single Policy is Near-Optimal for Reward-Free Reinforcement Learning

no code implementations31 May 2022 Pedro Cisneros-Velarde, Boxiang Lyu, Sanmi Koyejo, Mladen Kolar

Although parallelism has been extensively used in reinforcement learning (RL), the quantitative effects of parallel exploration are not well understood theoretically.

Reinforcement Learning (RL)

Pessimism meets VCG: Learning Dynamic Mechanism Design via Offline Reinforcement Learning

no code implementations5 May 2022 Boxiang Lyu, Zhaoran Wang, Mladen Kolar, Zhuoran Yang

In the setting where the function approximation is employed to handle large state spaces, with only mild assumptions on the expressiveness of the function class, we are able to design a dynamic mechanism using offline reinforcement learning algorithms.

Offline RL reinforcement-learning +1

Personalized Federated Learning with Multiple Known Clusters

1 code implementation28 Apr 2022 Boxiang Lyu, Filip Hanzely, Mladen Kolar

We consider the problem of personalized federated learning when there are known cluster structures within users.

Personalized Federated Learning

L-SVRG and L-Katyusha with Adaptive Sampling

no code implementations31 Jan 2022 Boxin Zhao, Boxiang Lyu, Mladen Kolar

Stochastic gradient-based optimization methods, such as L-SVRG and its accelerated variant L-Katyusha (Kovalev et al., 2020), are widely used to train machine learning models. The theoretical and empirical performance of L-SVRG and L-Katyusha can be improved by sampling observations from a non-uniform distribution (Qian et al., 2021).

Cannot find the paper you are looking for? You can Submit a new open access paper.