Search Results for author: Sedrick Keh

Found 8 papers, 6 papers with code

Should VLMs be Pre-trained with Image Data?

no code implementations10 Mar 2025 Sedrick Keh, Jean Mercat, Samir Yitzhak Gadre, Kushal Arora, Igor Vasiljevic, Benjamin Burchfiel, Shuran Song, Russ Tedrake, Thomas Kollar, Ludwig Schmidt, Achal Dave

We find that pre-training with a mixture of image and text data allows models to perform better on vision-language tasks while maintaining strong performance on text-only evaluations.

Linearizing Large Language Models

1 code implementation10 May 2024 Jean Mercat, Igor Vasiljevic, Sedrick Keh, Kushal Arora, Achal Dave, Adrien Gaidon, Thomas Kollar

Linear transformers have emerged as a subquadratic-time alternative to softmax attention and have garnered significant interest due to their fixed-size recurrent state that lowers inference cost.

In-Context Learning Mamba

A Critical Evaluation of AI Feedback for Aligning Large Language Models

1 code implementation19 Feb 2024 Archit Sharma, Sedrick Keh, Eric Mitchell, Chelsea Finn, Kushal Arora, Thomas Kollar

RLAIF first performs supervised fine-tuning (SFT) using demonstrations from a teacher model and then further fine-tunes the model with reinforcement learning (RL), using feedback from a critic model.

Instruction Following reinforcement-learning +2

Asking More Informative Questions for Grounded Retrieval

no code implementations14 Nov 2023 Sedrick Keh, Justin T. Chiu, Daniel Fried

When a model is trying to gather information in an interactive setting, it benefits from asking informative questions.

Question Answering Question Selection +2

Cannot find the paper you are looking for? You can Submit a new open access paper.