Search Results for author: Gopala Anumanchipalli

Found 8 papers, 2 papers with code

A Unified Framework for Model Editing

2 code implementations21 Mar 2024 Akshat Gupta, Dev Sajnani, Gopala Anumanchipalli

We introduce a unifying framework that brings two leading "locate-and-edit" model editing techniques -- ROME and MEMIT -- under a single conceptual umbrella, optimizing for the same goal, which we call the preservation-memorization objective.

Memorization Model Editing

Rebuilding ROME : Resolving Model Collapse during Sequential Model Editing

1 code implementation11 Mar 2024 Akshat Gupta, Sidharth Baskaran, Gopala Anumanchipalli

With this paper, we provide a more stable implementation ROME, which we call r-ROME and show that model collapse is no longer observed when making large scale sequential edits with r-ROME, while further improving generalization and locality of model editing compared to the original implementation of ROME.

Model Editing

Identifying Multiple Personalities in Large Language Models with External Evaluation

no code implementations22 Feb 2024 Xiaoyang Song, Yuta Adachi, Jessie Feng, Mouwei Lin, Linhao Yu, Frank Li, Akshat Gupta, Gopala Anumanchipalli, Simerjot Kaur

In this paper, we investigate LLM personalities using an alternate personality measurement method, which we refer to as the external evaluation method, where instead of prompting LLMs with multiple-choice questions in the Likert scale, we evaluate LLMs' personalities by analyzing their responses toward open-ended situational questions using an external machine learning model.

Multiple-choice

Towards Hierarchical Spoken Language Dysfluency Modeling

no code implementations18 Jan 2024 Jiachen Lian, Gopala Anumanchipalli

Speech disfluency modeling is the bottleneck for both speech therapy and language learning.

Model Editing at Scale leads to Gradual and Catastrophic Forgetting

no code implementations15 Jan 2024 Akshat Gupta, Anurag Rao, Gopala Anumanchipalli

With this in mind, we evaluate the current model editing methods at scale, focusing on two state of the art methods: ROME and MEMIT.

Model Editing Specificity

Self-Assessment Tests are Unreliable Measures of LLM Personality

no code implementations15 Sep 2023 Akshat Gupta, Xiaoyang Song, Gopala Anumanchipalli

These simple tests, done on ChatGPT and three Llama2 models of different sizes, show that self-assessment personality tests created for humans are unreliable measures of personality in LLMs.

Multiple-choice

Cannot find the paper you are looking for? You can Submit a new open access paper.