Search Results for author: Joe Edelman

Found 1 papers, 0 papers with code

What are human values, and how do we align AI to them?

no code implementations27 Mar 2024 Oliver Klingefjord, Ryan Lowe, Joe Edelman

In this paper, we focus on the first two parts, and ask the question: what are "good" ways to synthesize diverse human inputs about values into a target for aligning language models?

Language Modelling Large Language Model +1

Cannot find the paper you are looking for? You can Submit a new open access paper.