Search Results for author: Kanishk Gandhi

Found 11 papers, 4 papers with code

Self-Supervised Alignment with Mutual Information: Learning to Follow Principles without Preference Labels

1 code implementation22 Apr 2024 Jan-Philipp Fränken, Eric Zelikman, Rafael Rafailov, Kanishk Gandhi, Tobias Gerstenberg, Noah D. Goodman

On single-turn dialogue and summarization, a SAMI-trained mistral-7b outperforms the initial pretrained model, with win rates between 66% and 77%.

Language Modelling

Procedural Dilemma Generation for Evaluating Moral Reasoning in Humans and Language Models

1 code implementation17 Apr 2024 Jan-Philipp Fränken, Kanishk Gandhi, Tori Qiu, Ayesha Khawaja, Noah D. Goodman, Tobias Gerstenberg

We collected moral permissibility and intention judgments from human participants for a subset of our items and compared these judgments to those from two language models (GPT-4 and Claude-2) across eight conditions.

Decision Making Language Modelling +1

Stream of Search (SoS): Learning to Search in Language

1 code implementation1 Apr 2024 Kanishk Gandhi, Denise Lee, Gabriel Grand, Muxin Liu, Winson Cheng, Archit Sharma, Noah D. Goodman

In this paper, we show how language models can be taught to search by representing the process of search in language, as a flattened string -- a stream of search (SoS).

Language Modelling

Social Contract AI: Aligning AI Assistants with Implicit Group Norms

1 code implementation26 Oct 2023 Jan-Philipp Fränken, Sam Kwok, Peixuan Ye, Kanishk Gandhi, Dilip Arumugam, Jared Moore, Alex Tamkin, Tobias Gerstenberg, Noah D. Goodman

We explore the idea of aligning an AI assistant by inverting a model of users' (unknown) preferences from observed interactions.

Exciton-Polariton Condensates: A Fourier Neural Operator Approach

no code implementations27 Sep 2023 Surya T. Sathujoda, YuAn Wang, Kanishk Gandhi

Advancements in semiconductor fabrication over the past decade have catalyzed extensive research into all-optical devices driven by exciton-polariton condensates.

Understanding Social Reasoning in Language Models with Language Models

no code implementations NeurIPS 2023 Kanishk Gandhi, Jan-Philipp Fränken, Tobias Gerstenberg, Noah D. Goodman

Using our framework, we create a new social reasoning benchmark (BigToM) for LLMs which consists of 25 controls and 5, 000 model-written evaluations.

Certified Deductive Reasoning with Language Models

no code implementations6 Jun 2023 Gabriel Poesia, Kanishk Gandhi, Eric Zelikman, Noah D. Goodman

In experiments on PrOntoQA, ProofWriter and Syllogism Validity datasets, \textsc{LogicGuide} significantly improves the performance of GPT-3, GPT-3. 5 Turbo and LLaMA (accuracy gains up to 35\%), while drastically reducing \emph{content effects} -- the interference between unwanted prior assumptions and reasoning, which humans and language models suffer from.

Logical Reasoning valid

Strategic Reasoning with Language Models

no code implementations30 May 2023 Kanishk Gandhi, Dorsa Sadigh, Noah D. Goodman

Existing approaches to solving strategic games rely on extensive training, yielding strategies that do not generalize to new scenarios or games without retraining.

Eliciting Compatible Demonstrations for Multi-Human Imitation Learning

no code implementations14 Oct 2022 Kanishk Gandhi, Siddharth Karamcheti, Madeline Liao, Dorsa Sadigh

Imitation learning from human-provided demonstrations is a strong approach for learning policies for robot manipulation.

Imitation Learning Robot Manipulation

Baby Intuitions Benchmark (BIB): Discerning the goals, preferences, and actions of others

no code implementations NeurIPS 2021 Kanishk Gandhi, Gala Stojnic, Brenden M. Lake, Moira R. Dillon

To achieve human-like common sense about everyday life, machine learning systems must understand and reason about the goals, preferences, and actions of other agents in the environment.

Common Sense Reasoning

Cannot find the paper you are looking for? You can Submit a new open access paper.