Search Results for author: Nathan Lambert

Found 19 papers, 11 papers with code

Social Choice for AI Alignment: Dealing with Diverse Human Feedback

no code implementations16 Apr 2024 Vincent Conitzer, Rachel Freedman, Jobst Heitzig, Wesley H. Holliday, Bob M. Jacobs, Nathan Lambert, Milan Mossé, Eric Pacuit, Stuart Russell, Hailey Schoelkopf, Emanuel Tewolde, William S. Zwicker

Foundation models such as GPT-4 are fine-tuned to avoid unsafe or otherwise problematic behavior, so that, for example, they refuse to comply with requests for help with committing crimes or with producing racist text.

Ethics

Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2

2 code implementations17 Nov 2023 Hamish Ivison, Yizhong Wang, Valentina Pyatkin, Nathan Lambert, Matthew Peters, Pradeep Dasigi, Joel Jang, David Wadden, Noah A. Smith, Iz Beltagy, Hannaneh Hajishirzi

Since the release of T\"ULU [Wang et al., 2023b], open resources for instruction tuning have developed quickly, from better base models to new finetuning techniques.

The Alignment Ceiling: Objective Mismatch in Reinforcement Learning from Human Feedback

no code implementations31 Oct 2023 Nathan Lambert, Roberto Calandra

Reinforcement learning from human feedback (RLHF) has emerged as a powerful technique to make large language models (LLMs) more capable in complex settings.

GSM8K Model-based Reinforcement Learning +1

Zephyr: Direct Distillation of LM Alignment

1 code implementation25 Oct 2023 Lewis Tunstall, Edward Beeching, Nathan Lambert, Nazneen Rajani, Kashif Rasul, Younes Belkada, Shengyi Huang, Leandro von Werra, Clémentine Fourrier, Nathan Habib, Nathan Sarrazin, Omar Sanseviero, Alexander M. Rush, Thomas Wolf

Starting from a dataset of outputs ranked by a teacher model, we apply distilled direct preference optimization (dDPO) to learn a chat model with significantly improved intent alignment.

2D Cyclist Detection Language Modelling

A Unified View on Solving Objective Mismatch in Model-Based Reinforcement Learning

1 code implementation10 Oct 2023 Ran Wei, Nathan Lambert, Anthony McDonald, Alfredo Garcia, Roberto Calandra

Model-based Reinforcement Learning (MBRL) aims to make agents more sample-efficient, adaptive, and explainable by learning an explicit model of the environment.

Model-based Reinforcement Learning

BLISS: Interplanetary Exploration with Swarms of Low-Cost Spacecraft

no code implementations20 Jul 2023 Alexander N. Alvara, Lydia Lee, Emmanuel Sin, Nathan Lambert, Andrew J. Westphal, Kristofer S. J. Pister

Leveraging advancements in micro-scale technology, we propose a fleet of autonomous, low-cost, small solar sails for interplanetary exploration.

Measuring Data

no code implementations9 Dec 2022 Margaret Mitchell, Alexandra Sasha Luccioni, Nathan Lambert, Marissa Gerchick, Angelina McMillan-Major, Ezinwanne Ozoani, Nazneen Rajani, Tristan Thrush, Yacine Jernite, Douwe Kiela

We identify the task of measuring data to quantitatively characterize the composition of machine learning data and datasets.

Reward Reports for Reinforcement Learning

1 code implementation22 Apr 2022 Thomas Krendl Gilbert, Nathan Lambert, Sarah Dean, Tom Zick, Aaron Snoswell

Building systems that are good for society in the face of complex societal effects requires a dynamic approach.

Chatbot reinforcement-learning +1

Investigating Compounding Prediction Errors in Learned Dynamics Models

no code implementations17 Mar 2022 Nathan Lambert, Kristofer Pister, Roberto Calandra

In this paper, we explore the effects of subcomponents of a control problem on long term prediction error: including choosing a system, collecting data, and training a model.

Model-based Reinforcement Learning

Choices, Risks, and Reward Reports: Charting Public Policy for Reinforcement Learning Systems

1 code implementation11 Feb 2022 Thomas Krendl Gilbert, Sarah Dean, Tom Zick, Nathan Lambert

In the long term, reinforcement learning (RL) is considered by many AI theorists to be the most promising path to artificial general intelligence.

Recommendation Systems reinforcement-learning +1

The Challenges of Exploration for Offline Reinforcement Learning

no code implementations27 Jan 2022 Nathan Lambert, Markus Wulfmeier, William Whitney, Arunkumar Byravan, Michael Bloesch, Vibhavari Dasagi, Tim Hertweck, Martin Riedmiller

Offline Reinforcement Learning (ORL) enablesus to separately study the two interlinked processes of reinforcement learning: collecting informative experience and inferring optimal behaviour.

Model Predictive Control Offline RL +2

On the Importance of Hyperparameter Optimization for Model-based Reinforcement Learning

1 code implementation26 Feb 2021 Baohe Zhang, Raghu Rajan, Luis Pineda, Nathan Lambert, André Biedenkapp, Kurtland Chua, Frank Hutter, Roberto Calandra

We demonstrate that this problem can be tackled effectively with automated HPO, which we demonstrate to yield significantly improved performance compared to human experts.

Hyperparameter Optimization Model-based Reinforcement Learning +2

AI Development for the Public Interest: From Abstraction Traps to Sociotechnical Risks

no code implementations4 Feb 2021 McKane Andrus, Sarah Dean, Thomas Krendl Gilbert, Nathan Lambert, Tom Zick

Despite interest in communicating ethical problems and social contexts within the undergraduate curriculum to advance Public Interest Technology (PIT) goals, interventions at the graduate level remain largely unexplored.

Nonholonomic Yaw Control of an Underactuated Flying Robot with Model-based Reinforcement Learning

no code implementations2 Sep 2020 Nathan Lambert, Craig Schindler, Daniel Drew, Kristofer Pister

As a comparison to the significant engineering effort required for an analytic control law, we implement a data-driven model-based reinforcement learning yaw controller in a simulated flight task.

Robotics

Objective Mismatch in Model-based Reinforcement Learning

2 code implementations ICLR 2020 Nathan Lambert, Brandon Amos, Omry Yadan, Roberto Calandra

In our experiments, we study this objective mismatch issue and demonstrate that the likelihood of one-step ahead predictions is not always correlated with control performance.

Model-based Reinforcement Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.