no code implementations • 12 Mar 2025 • Lutfi Eren Erdogan, Nicholas Lee, Sehoon Kim, Suhong Moon, Hiroki Furuta, Gopala Anumanchipalli, Kurt Keutzer, Amir Gholami
Large language models (LLMs) have shown remarkable advancements in enabling language agents to tackle simple tasks.
1 code implementation • 6 Mar 2025 • Guan-Ting Lin, Jiachen Lian, Tingle Li, Qirui Wang, Gopala Anumanchipalli, Alexander H. Liu, Hung-Yi Lee
Spoken dialogue modeling introduces unique challenges beyond text-based language modeling, demanding robust turn-taking, backchanneling, and real-time interaction.
no code implementations • 21 Jan 2025 • Kan Jen Cheng, Tingle Li, Gopala Anumanchipalli
Audio texture manipulation involves modifying the perceptual characteristics of a sound to achieve specific transformations, such as adding, removing, or replacing auditory elements.
1 code implementation • 14 Jan 2025 • Richard Zhuang, Akshat Gupta, Richard Yang, Aniket Rahane, Zhengyu Li, Gopala Anumanchipalli
PokerBench thus presents a unique benchmark for a quick and reliable evaluation of the poker-playing ability of LLMs as well as a comprehensive benchmark to study the progress of LLMs in complex game-playing scenarios.
1 code implementation • 25 Sep 2024 • Fazal Mittu, Yihuan Bu, Akshat Gupta, Ashok Devireddy, Alp Eren Ozdarendeli, Anant Singh, Gopala Anumanchipalli
We compare traditional text compression systems with neural network and LLM-based text compression methods.
no code implementations • 22 Sep 2024 • Tingle Li, Renhao Wang, Po-Yao Huang, Andrew Owens, Gopala Anumanchipalli
Through this process, the model learns to transfer the conditional example's sound properties to the input speech.
no code implementations • 20 Sep 2024 • Xuanru Zhou, Jiachen Lian, Cheol Jun Cho, Jingwen Liu, Zongli Ye, Jinming Zhang, Brittany Morin, David Baquirin, Jet Vonk, Zoe Ezzes, Zachary Miller, Maria Luisa Gorno Tempini, Gopala Anumanchipalli
In this work, we revisit this problem from a new perspective: tokenizing dysfluencies and modeling the detection problem as a token-based automatic speech recognition (ASR) problem.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
no code implementations • 19 Sep 2024 • Akshat Gupta, Atahan Ozdemir, Gopala Anumanchipalli
Finally, we compare the hidden representations of LayerNorm-based LLMs with models trained using RMSNorm and show that all LLMs naturally operate orthogonal to the uniform vector at inference time, that is, on average they do not have a component along the uniform vector during inference.
no code implementations • 15 Sep 2024 • Xuanru Zhou, Cheol Jun Cho, Ayati Sharma, Brittany Morin, David Baquirin, Jet Vonk, Zoe Ezzes, Zachary Miller, Boon Lead Tee, Maria Luisa Gorno Tempini, Jiachen Lian, Gopala Anumanchipalli
Current de-facto dysfluency modeling methods utilize template matching algorithms which are not generalizable to out-of-domain real-world dysfluencies across languages, and are not scalable with increasing amounts of training data.
1 code implementation • 1 Sep 2024 • Lutfi Eren Erdogan, Nicholas Lee, Siddharth Jha, Sehoon Kim, Ryan Tabrizi, Suhong Moon, Coleman Hooper, Gopala Anumanchipalli, Kurt Keutzer, Amir Gholami
Recent large language models (LLMs) have enabled the development of advanced agentic systems that can integrate various tools and APIs to fulfill user queries through function calling.
no code implementations • 22 Jun 2024 • Rishi Jain, Bohan Yu, Peter Wu, Tejas Prabhune, Gopala Anumanchipalli
Together, we set a new benchmark for vocal tract modeling in MRI video segmentation and use this to release labels for a 75-speaker RT-MRI dataset, increasing the amount of labeled public RT-MRI data of the vocal tract by over a factor of 9.
1 code implementation • 1 May 2024 • Junsang Yoon, Akshat Gupta, Gopala Anumanchipalli
This study presents a targeted model editing analysis focused on the latest large language model, Llama-3.
1 code implementation • 22 Mar 2024 • Nicholas Lee, Thanakul Wattanawong, Sehoon Kim, Karttikeya Mangalam, Sheng Shen, Gopala Anumanchipalli, Michael W. Mahoney, Kurt Keutzer, Amir Gholami
To address this, we propose LLM2LLM, a targeted and iterative data augmentation strategy that uses a teacher LLM to enhance a small seed dataset by augmenting additional data that can be used for fine-tuning on a specific task.
1 code implementation • 21 Mar 2024 • Akshat Gupta, Dev Sajnani, Gopala Anumanchipalli
We generalize ROME and enable batched editing with equality constraint in the form of EMMET - an Equality-constrained Mass Model Editing algorithm for Transformers, a new batched memory-editing algorithm.
1 code implementation • 11 Mar 2024 • Akshat Gupta, Sidharth Baskaran, Gopala Anumanchipalli
With this paper, we provide a more stable implementation ROME, which we call r-ROME and show that model collapse is no longer observed when making large scale sequential edits with r-ROME, while further improving generalization and locality of model editing compared to the original implementation of ROME.
no code implementations • 22 Feb 2024 • Xiaoyang Song, Yuta Adachi, Jessie Feng, Mouwei Lin, Linhao Yu, Frank Li, Akshat Gupta, Gopala Anumanchipalli, Simerjot Kaur
In this paper, we investigate LLM personalities using an alternate personality measurement method, which we refer to as the external evaluation method, where instead of prompting LLMs with multiple-choice questions in the Likert scale, we evaluate LLMs' personalities by analyzing their responses toward open-ended situational questions using an external machine learning model.
no code implementations • 18 Jan 2024 • Jiachen Lian, Gopala Anumanchipalli
Speech disfluency modeling is the bottleneck for both speech therapy and language learning.
1 code implementation • 15 Jan 2024 • Akshat Gupta, Anurag Rao, Gopala Anumanchipalli
With this in mind, we evaluate the current model editing methods at scale, focusing on two state of the art methods: ROME and MEMIT.
no code implementations • 4 Oct 2023 • Robin Netzorg, Bohan Yu, Andrea Guzman, Peter Wu, Luna McNulty, Gopala Anumanchipalli
Unlike other data modalities such as text and vision, speech does not lend itself to easy interpretation.
no code implementations • 15 Sep 2023 • Akshat Gupta, Xiaoyang Song, Gopala Anumanchipalli
These simple tests, done on ChatGPT and three Llama2 models of different sizes, show that self-assessment personality tests created for humans are unreliable measures of personality in LLMs.