Search Results for author: Zhaolin Gao

Found 3 papers, 1 papers with code

REBEL: Reinforcement Learning via Regressing Relative Rewards

no code implementations25 Apr 2024 Zhaolin Gao, Jonathan D. Chang, Wenhao Zhan, Owen Oertell, Gokul Swamy, Kianté Brantley, Thorsten Joachims, J. Andrew Bagnell, Jason D. Lee, Wen Sun

While originally developed for continuous control problems, Proximal Policy Optimization (PPO) has emerged as the work-horse of a variety of reinforcement learning (RL) applications including the fine-tuning of generative models.

Continuous Control Image Generation +3

Reviewer2: Optimizing Review Generation Through Prompt Generation

no code implementations16 Feb 2024 Zhaolin Gao, Kianté Brantley, Thorsten Joachims

In this paper, we envision a use case where authors can receive LLM-generated reviews that uncover weak points in the current draft.

Review Generation

Shoestring: Graph-Based Semi-Supervised Learning with Severely Limited Labeled Data

1 code implementation28 Oct 2019 Wan-Yu Lin, Zhaolin Gao, Baochun Li

More specifically, we address the problem of graph-based semi-supervised learning in the presence of severely limited labeled samples, and propose a new framework, called {\em Shoestring}, that improves the learning performance through semantic transfer from these very few labeled samples to large numbers of unlabeled samples.

Few-Shot Learning General Classification +4

Cannot find the paper you are looking for? You can Submit a new open access paper.