Search Results for author: Gopala Anumanchipalli

Found 20 papers, 9 papers with code

Full-Duplex-Bench: A Benchmark to Evaluate Full-duplex Spoken Dialogue Models on Turn-taking Capabilities

1 code implementation6 Mar 2025 Guan-Ting Lin, Jiachen Lian, Tingle Li, Qirui Wang, Gopala Anumanchipalli, Alexander H. Liu, Hung-Yi Lee

Spoken dialogue modeling introduces unique challenges beyond text-based language modeling, demanding robust turn-taking, backchanneling, and real-time interaction.

Language Modeling Language Modelling +1

Audio Texture Manipulation by Exemplar-Based Analogy

no code implementations21 Jan 2025 Kan Jen Cheng, Tingle Li, Gopala Anumanchipalli

Audio texture manipulation involves modifying the perceptual characteristics of a sound to achieve specific transformations, such as adding, removing, or replacing auditory elements.

PokerBench: Training Large Language Models to become Professional Poker Players

1 code implementation14 Jan 2025 Richard Zhuang, Akshat Gupta, Richard Yang, Aniket Rahane, Zhengyu Li, Gopala Anumanchipalli

PokerBench thus presents a unique benchmark for a quick and reliable evaluation of the poker-playing ability of LLMs as well as a comprehensive benchmark to study the progress of LLMs in complex game-playing scenarios.

Self-Supervised Audio-Visual Soundscape Stylization

no code implementations22 Sep 2024 Tingle Li, Renhao Wang, Po-Yao Huang, Andrew Owens, Gopala Anumanchipalli

Through this process, the model learns to transfer the conditional example's sound properties to the input speech.

Speech Enhancement

Geometric Interpretation of Layer Normalization and a Comparative Analysis with RMSNorm

no code implementations19 Sep 2024 Akshat Gupta, Atahan Ozdemir, Gopala Anumanchipalli

Finally, we compare the hidden representations of LayerNorm-based LLMs with models trained using RMSNorm and show that all LLMs naturally operate orthogonal to the uniform vector at inference time, that is, on average they do not have a component along the uniform vector during inference.

Stutter-Solver: End-to-end Multi-lingual Dysfluency Detection

no code implementations15 Sep 2024 Xuanru Zhou, Cheol Jun Cho, Ayati Sharma, Brittany Morin, David Baquirin, Jet Vonk, Zoe Ezzes, Zachary Miller, Boon Lead Tee, Maria Luisa Gorno Tempini, Jiachen Lian, Gopala Anumanchipalli

Current de-facto dysfluency modeling methods utilize template matching algorithms which are not generalizable to out-of-domain real-world dysfluencies across languages, and are not scalable with increasing amounts of training data.

object-detection Object Detection +1

TinyAgent: Function Calling at the Edge

1 code implementation1 Sep 2024 Lutfi Eren Erdogan, Nicholas Lee, Siddharth Jha, Sehoon Kim, Ryan Tabrizi, Suhong Moon, Coleman Hooper, Gopala Anumanchipalli, Kurt Keutzer, Amir Gholami

Recent large language models (LLMs) have enabled the development of advanced agentic systems that can integrate various tools and APIs to fulfill user queries through function calling.

Language Modelling Quantization +1

Multimodal Segmentation for Vocal Tract Modeling

no code implementations22 Jun 2024 Rishi Jain, Bohan Yu, Peter Wu, Tejas Prabhune, Gopala Anumanchipalli

Together, we set a new benchmark for vocal tract modeling in MRI video segmentation and use this to release labels for a 75-speaker RT-MRI dataset, increasing the amount of labeled public RT-MRI data of the vocal tract by over a factor of 9.

Segmentation Video Segmentation +1

LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement

1 code implementation22 Mar 2024 Nicholas Lee, Thanakul Wattanawong, Sehoon Kim, Karttikeya Mangalam, Sheng Shen, Gopala Anumanchipalli, Michael W. Mahoney, Kurt Keutzer, Amir Gholami

To address this, we propose LLM2LLM, a targeted and iterative data augmentation strategy that uses a teacher LLM to enhance a small seed dataset by augmenting additional data that can be used for fine-tuning on a specific task.

Data Augmentation GSM8K +1

A Unified Framework for Model Editing

1 code implementation21 Mar 2024 Akshat Gupta, Dev Sajnani, Gopala Anumanchipalli

We generalize ROME and enable batched editing with equality constraint in the form of EMMET - an Equality-constrained Mass Model Editing algorithm for Transformers, a new batched memory-editing algorithm.

Memorization model +1

Rebuilding ROME : Resolving Model Collapse during Sequential Model Editing

1 code implementation11 Mar 2024 Akshat Gupta, Sidharth Baskaran, Gopala Anumanchipalli

With this paper, we provide a more stable implementation ROME, which we call r-ROME and show that model collapse is no longer observed when making large scale sequential edits with r-ROME, while further improving generalization and locality of model editing compared to the original implementation of ROME.

model Model Editing

Identifying Multiple Personalities in Large Language Models with External Evaluation

no code implementations22 Feb 2024 Xiaoyang Song, Yuta Adachi, Jessie Feng, Mouwei Lin, Linhao Yu, Frank Li, Akshat Gupta, Gopala Anumanchipalli, Simerjot Kaur

In this paper, we investigate LLM personalities using an alternate personality measurement method, which we refer to as the external evaluation method, where instead of prompting LLMs with multiple-choice questions in the Likert scale, we evaluate LLMs' personalities by analyzing their responses toward open-ended situational questions using an external machine learning model.

Multiple-choice

Towards Hierarchical Spoken Language Dysfluency Modeling

no code implementations18 Jan 2024 Jiachen Lian, Gopala Anumanchipalli

Speech disfluency modeling is the bottleneck for both speech therapy and language learning.

Model Editing at Scale leads to Gradual and Catastrophic Forgetting

1 code implementation15 Jan 2024 Akshat Gupta, Anurag Rao, Gopala Anumanchipalli

With this in mind, we evaluate the current model editing methods at scale, focusing on two state of the art methods: ROME and MEMIT.

Model Editing Specificity

Self-Assessment Tests are Unreliable Measures of LLM Personality

no code implementations15 Sep 2023 Akshat Gupta, Xiaoyang Song, Gopala Anumanchipalli

These simple tests, done on ChatGPT and three Llama2 models of different sizes, show that self-assessment personality tests created for humans are unreliable measures of personality in LLMs.

Multiple-choice

Cannot find the paper you are looking for? You can Submit a new open access paper.