no code implementations • SIGDIAL (ACL) 2021 • Satwik Kottur, Paul Crook, Seungwhan Moon, Ahmad Beirami, Eunjoon Cho, Rajen Subba, Alborz Geramifard
There is a growing interest in virtual assistants with multimodal capabilities, e. g., inferring the context of a conversation through scene understanding.
no code implementations • 24 Jun 2024 • James Atwood, Preethi Lahoti, Ananth Balashankar, Flavien Prost, Ahmad Beirami
Prompting Large Language Models (LLMs) has created new and interesting means for classifying textual data.
1 code implementation • 10 Jun 2024 • Xiangyu Qi, Ashwinee Panda, Kaifeng Lyu, Xiao Ma, Subhrajit Roy, Ahmad Beirami, Prateek Mittal, Peter Henderson
We refer to this issue as shallow safety alignment.
no code implementations • 29 May 2024 • Adam Fisch, Jacob Eisenstein, Vicky Zayats, Alekh Agarwal, Ahmad Beirami, Chirag Nagpal, Pete Shaw, Jonathan Berant
Moreover, to account for uncertainty in the reward model we are distilling from, we optimize against a family of reward models that, as a whole, is likely to include at least one reasonable proxy for the preference distribution.
no code implementations • 28 May 2024 • Pritam Sarkar, Sayna Ebrahimi, Ali Etemad, Ahmad Beirami, Sercan Ö. Arik, Tomas Pfister
For a given factual token, we create a hallucinated token through generative data augmentation by selectively altering the ground-truth information.
no code implementations • 18 Apr 2024 • Zhaofeng Wu, Ananth Balashankar, Yoon Kim, Jacob Eisenstein, Ahmad Beirami
In this work, we evaluate a simple approach for zero-shot cross-lingual alignment, where a reward model is trained on preference data in one source language and directly applied to other target languages.
no code implementations • 2 Apr 2024 • Joy Qiping Yang, Salman Salamatian, Ziteng Sun, Ananda Theertha Suresh, Ahmad Beirami
The goal of language model alignment is to alter $p$ to a new distribution $\phi$ that results in a higher expected reward while keeping $\phi$ close to $p.$ A popular alignment method is the KL-constrained reinforcement learning (RL), which chooses a distribution $\phi_\Delta$ that maximizes $E_{\phi_{\Delta}} r(y)$ subject to a relative entropy constraint $KL(\phi_\Delta || p) \leq \Delta.$ Another simple alignment method is best-of-$N$, where $N$ samples are drawn from $p$ and one with highest reward is selected.
no code implementations • 15 Mar 2024 • Ziteng Sun, Uri Mendlovic, Yaniv Leviathan, Asaf Aharoni, Ahmad Beirami, Jae Hun Ro, Ananda Theertha Suresh
Speculative decoding is an effective method for lossless acceleration of large language models during inference.
1 code implementation • 30 Jan 2024 • Nevan Wichers, Carson Denison, Ahmad Beirami
Red teaming is a common strategy for identifying weaknesses in generative language models (LMs), where adversarial prompts are produced that trigger an LM to generate unsafe responses.
no code implementations • 3 Jan 2024 • Ahmad Beirami, Alekh Agarwal, Jonathan Berant, Alexander D'Amour, Jacob Eisenstein, Chirag Nagpal, Ananda Theertha Suresh
A commonly used analytical expression in the literature claims that the KL divergence between the best-of-$n$ policy and the base policy is equal to $\log (n) - (n-1)/n.$ We disprove the validity of this claim, and show that it is an upper bound on the actual KL divergence.
1 code implementation • 14 Dec 2023 • Jacob Eisenstein, Chirag Nagpal, Alekh Agarwal, Ahmad Beirami, Alex D'Amour, DJ Dvijotham, Adam Fisch, Katherine Heller, Stephen Pfohl, Deepak Ramachandran, Peter Shaw, Jonathan Berant
However, even pretrain reward ensembles do not eliminate reward hacking: we show several qualitative reward hacking phenomena that are not mitigated by ensembling because all reward models in the ensemble exhibit similar error patterns.
no code implementations • 6 Dec 2023 • Lucas Monteiro Paes, Ananda Theertha Suresh, Alex Beutel, Flavio P. Calmon, Ahmad Beirami
Here, the sample complexity for estimating the worst-case performance gap across groups (e. g., the largest difference in error rates) increases exponentially with the number of group-denoting sensitive attributes.
1 code implementation • 5 Dec 2023 • Alexandru Tifrea, Preethi Lahoti, Ben Packer, Yoni Halpern, Ahmad Beirami, Flavien Prost
Despite achieving promising fairness-error trade-offs, in-processing mitigation techniques for group fairness cannot be employed in numerous practical applications with limited computation resources or no access to the training pipeline of the prediction model.
1 code implementation • 2 Nov 2023 • Bhagyashree Puranik, Ahmad Beirami, Yao Qin, Upamanyu Madhow
State-of-the-art techniques for enhancing robustness of deep networks mostly rely on empirical risk minimization with suitable data augmentation.
no code implementations • 25 Oct 2023 • Sidharth Mudgal, Jong Lee, Harish Ganapathy, Yaguang Li, Tao Wang, Yanping Huang, Zhifeng Chen, Heng-Tze Cheng, Michael Collins, Trevor Strohman, Jilin Chen, Alex Beutel, Ahmad Beirami
KL-regularized reinforcement learning (RL) is a popular alignment framework to control the language model responses towards high reward outcomes.
Language Modelling Multi-Objective Reinforcement Learning +1
no code implementations • 25 Oct 2023 • Aradhana Sinha, Ananth Balashankar, Ahmad Beirami, Thi Avrahami, Jilin Chen, Alex Beutel
We demonstrate the advantages of this system on the ANLI and hate speech detection benchmark datasets - both collected via an iterative, adversarial human-and-model-in-the-loop procedure.
no code implementations • 25 Oct 2023 • Ananth Balashankar, Xiao Ma, Aradhana Sinha, Ahmad Beirami, Yao Qin, Jilin Chen, Alex Beutel
As large language models (LLMs) are widely adopted, new safety issues and policies emerge, to which existing safety classifiers do not generalize well.
no code implementations • 25 Oct 2023 • Preethi Lahoti, Nicholas Blumm, Xiao Ma, Raghavendra Kotikalapudi, Sahitya Potluri, Qijun Tan, Hansa Srinivasan, Ben Packer, Ahmad Beirami, Alex Beutel, Jilin Chen
A crucial challenge for generative large language models (LLMs) is diversity: when a user's prompt is under-specified, models may follow implicit assumptions while generating a response, which may result in homogenization of the responses, as well as certain demographic groups being under-represented or even erased from the generated responses.
no code implementations • NeurIPS 2023 • Ziteng Sun, Ananda Theertha Suresh, Jae Hun Ro, Ahmad Beirami, Himanshu Jain, Felix Yu
We show that the optimal draft selection algorithm (transport plan) can be computed via linear programming, whose best-known runtime is exponential in $k$.
2 code implementations • 17 Oct 2023 • Somnath Basu Roy Chowdhury, Nicholas Monath, Ahmad Beirami, Rahul Kidambi, Avinava Dubey, Amr Ahmed, Snigdha Chaturvedi
In the online setting, where the algorithm has access to a single instance at a time, estimating the group fairness objective requires additional storage and significantly more computation (e. g., forward/backward passes) than the task-specific objective at every time step.
2 code implementations • 24 Jul 2023 • Jindong Gu, Zhen Han, Shuo Chen, Ahmad Beirami, Bailan He, Gengyuan Zhang, Ruotong Liao, Yao Qin, Volker Tresp, Philip Torr
This paper aims to provide a comprehensive survey of cutting-edge research in prompt engineering on three types of vision-language models: multimodal-to-text generation models (e. g. Flamingo), image-text matching models (e. g.
no code implementations • 11 Jul 2023 • James Atwood, Tina Tian, Ben Packer, Meghana Deodhar, Jilin Chen, Alex Beutel, Flavien Prost, Ahmad Beirami
Despite the rich literature on machine learning fairness, relatively little attention has been paid to remediating complex systems, where the final prediction is the combination of multiple classifiers and where multiple groups are present.
no code implementations • 25 Jun 2023 • Xiao Ma, Swaroop Mishra, Ahmad Beirami, Alex Beutel, Jilin Chen
Language models still struggle on moral reasoning, despite their impressive performance in many other tasks.
1 code implementation • NeurIPS 2023 • Pritam Sarkar, Ahmad Beirami, Ali Etemad
Video self-supervised learning (VSSL) has made significant progress in recent years.
no code implementations • 17 Apr 2023 • Jindong Gu, Ahmad Beirami, Xuezhi Wang, Alex Beutel, Philip Torr, Yao Qin
With the advent of vision-language models (VLMs) that can perform in-context and prompt-based learning, how can we design prompting approaches that robustly generalize to distribution shift and can be used on novel classes outside the support set of the prompts?
1 code implementation • NAACL 2022 • Ninareh Mehrabi, Ahmad Beirami, Fred Morstatter, Aram Galstyan
Existing work to generate such attacks is either based on human-generated attacks which is costly and not scalable or, in case of automatic attacks, the attack vector does not conform to human-like language, which can be detected using a language model loss.
no code implementations • NAACL 2022 • Kun Qian, Ahmad Beirami, Satwik Kottur, Shahin Shayandeh, Paul Crook, Alborz Geramifard, Zhou Yu, Chinnadhurai Sankar
We find that training on our augmented dialog data improves the model's ability to deal with ambiguous scenarios, without sacrificing performance on unmodified turns.
2 code implementations • 15 Dec 2021 • Hyundong Cho, Chinnadhurai Sankar, Christopher Lin, Kaushik Ram Sadagopan, Shahin Shayandeh, Asli Celikyilmaz, Jonathan May, Ahmad Beirami
Recent works that revealed the vulnerability of dialogue state tracking (DST) models to distributional shifts have made holistic comparisons on robustness and qualitative analyses increasingly important for understanding their relative performance.
Ranked #4 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.1 (using extra training data)
Dialogue State Tracking Multi-domain Dialogue State Tracking +1
no code implementations • 8 Nov 2021 • Matthew Nokleby, Ahmad Beirami
For models that are (roughly) lower Lipschitz in their parameters, we bound the rate distortion function from below, whereas for VC classes, the mutual information is bounded above by $d_\mathrm{vc}\log(n)$.
no code implementations • 28 Oct 2021 • Nasser Aldaghri, Hessam Mahdavifar, Ahmad Beirami
We propose a new algorithm for FL with heterogeneous DP, named FedHDP, which employs personalization and weighted averaging at the server using the privacy choices of clients, to achieve better performance on clients' local models.
1 code implementation • 21 Oct 2021 • Tianjian Huang, Shaunak Halbe, Chinnadhurai Sankar, Pooyan Amini, Satwik Kottur, Alborz Geramifard, Meisam Razaviyayn, Ahmad Beirami
Our experiments show that DAIR consistently outperforms ERM and DA-ERM with little marginal computational cost and sets new state-of-the-art results in several benchmarks involving covariant data augmentation.
Ranked #1 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.2
Multi-domain Dialogue State Tracking Visual Question Answering
1 code implementation • 13 Sep 2021 • Tian Li, Ahmad Beirami, Maziar Sanjabi, Virginia Smith
Finally, we demonstrate that TERM can be used for a multitude of applications in machine learning, such as enforcing fairness between subgroups, mitigating the effect of outliers, and handling class imbalance.
no code implementations • SIGDIAL (ACL) 2021 • Kun Qian, Ahmad Beirami, Zhouhan Lin, Ankita De, Alborz Geramifard, Zhou Yu, Chinnadhurai Sankar
In this work, we identify an overlooked issue with dialog state annotation inconsistencies in the dataset, where a slot type is tagged inconsistently across similar dialogs leading to confusion for DST modeling.
1 code implementation • NeurIPS 2021 • Andrew Lowy, Sina Baharlouei, Rakesh Pavan, Meisam Razaviyayn, Ahmad Beirami
We consider the problem of fair classification with discrete sensitive attributes and potentially large models and data sets, requiring stochastic solvers.
1 code implementation • ACL 2021 • Hung Le, Chinnadhurai Sankar, Seungwhan Moon, Ahmad Beirami, Alborz Geramifard, Satwik Kottur
A video-grounded dialogue system is required to understand both dialogue, which contains semantic dependencies from turn to turn, and video, which contains visual cues of spatial and temporal scene variations.
no code implementations • 1 Jan 2021 • Rakesh Pavan, Andrew Lowy, Sina Baharlouei, Meisam Razaviyayn, Ahmad Beirami
In this paper, we propose another notion of fairness violation, called Exponential Rényi Mutual Information (ERMI) between sensitive attributes and the predicted target.
no code implementations • 31 Dec 2020 • Nasser Aldaghri, Hessam Mahdavifar, Ahmad Beirami
We also present the corresponding unlearning protocol and show that it satisfies the perfect unlearning criterion.
4 code implementations • 8 Dec 2020 • Tian Li, Shengyuan Hu, Ahmad Beirami, Virginia Smith
Fairness and robustness are two important concerns for federated learning systems.
no code implementations • 12 Nov 2020 • Chulaka Gunasekara, Seokhwan Kim, Luis Fernando D'Haro, Abhinav Rastogi, Yun-Nung Chen, Mihail Eric, Behnam Hedayatnia, Karthik Gopalakrishnan, Yang Liu, Chao-Wei Huang, Dilek Hakkani-Tür, Jinchao Li, Qi Zhu, Lingxiao Luo, Lars Liden, Kaili Huang, Shahin Shayandeh, Runze Liang, Baolin Peng, Zheng Zhang, Swadheen Shukla, Minlie Huang, Jianfeng Gao, Shikib Mehri, Yulan Feng, Carla Gordon, Seyed Hossein Alavi, David Traum, Maxine Eskenazi, Ahmad Beirami, Eunjoon, Cho, Paul A. Crook, Ankita De, Alborz Geramifard, Satwik Kottur, Seungwhan Moon, Shivani Poddar, Rajen Subba
Interactive evaluation of dialog, and 4.
no code implementations • COLING 2020 • Zhenpeng Zhou, Ahmad Beirami, Paul Crook, Pararth Shah, Rajen Subba, Alborz Geramifard
We explore the tasks of one-shot learning and zero-shot domain transfer with DILOG on SimDial and MultiWoZ.
2 code implementations • ICLR 2021 • Tian Li, Ahmad Beirami, Maziar Sanjabi, Virginia Smith
Empirical risk minimization (ERM) is typically designed to perform well on the average loss, which can result in estimators that are sensitive to outliers, generalize poorly, or treat subgroups unfairly.
no code implementations • 24 Jun 2020 • Sofia M Nikolakaki, Ogheneovo Dibie, Ahmad Beirami, Nicholas Peterson, Navid Aghdaie, Kazi Zaman
Competition is a primary driver of player satisfaction and engagement in multiplayer online games.
2 code implementations • COLING 2020 • Seungwhan Moon, Satwik Kottur, Paul A. Crook, Ankita De, Shivani Poddar, Theodore Levin, David Whitney, Daniel Difranco, Ahmad Beirami, Eunjoon Cho, Rajen Subba, Alborz Geramifard
Next generation virtual assistants are envisioned to handle multimodal inputs (e. g., vision, memories of previous interactions, in addition to the user's utterances), and perform multimodal actions (e. g., displaying a route in addition to generating the system's utterance).
no code implementations • ICLR 2020 • Sina Baharlouei, Maher Nouiehed, Ahmad Beirami, Meisam Razaviyayn
In this paper, we use R\'enyi correlation as a measure of fairness of machine learning models and develop a general training framework to impose fairness.
no code implementations • 25 Jun 2019 • Yunqi Zhao, Igor Borovikov, Jason Rupert, Caedmon Somers, Ahmad Beirami
In recent years, reinforcement learning has been successful in solving video games from Atari to Star Craft II.
2 code implementations • 3 Jun 2019 • Igor Borovikov, Jesse Harder, Michael Sadovsky, Ahmad Beirami
We propose to create such NPC behaviors interactively by training an agent in the target environment using imitation learning with a human in the loop.
2 code implementations • ICLR 2020 • Tian Li, Maziar Sanjabi, Ahmad Beirami, Virginia Smith
Federated learning involves training statistical models in massive, heterogeneous networks.
no code implementations • 25 Mar 2019 • Yunqi Zhao, Igor Borovikov, Fernando De Mesentier Silva, Ahmad Beirami, Jason Rupert, Caedmon Somers, Jesse Harder, John Kolen, Jervis Pinto, Reza Pourabolghasem, James Pestrak, Harold Chaput, Mohsen Sardari, Long Lin, Sundeep Narravula, Navid Aghdaie, Kazi Zaman
We discuss two fundamental metrics based on which we measure the human-likeness of agents, namely skill and style, which are multi-faceted concepts with practical implications outlined in this paper.
no code implementations • 26 Oct 2018 • Parinaz Farajiparvar, Ahmad Beirami, Matthew Nokleby
We consider this problem for unsupervised learning for batch and sequential data.
1 code implementation • 27 Jan 2018 • Ahmad Beirami, Robert Calderbank, Mark Christiansen, Ken Duffy, Muriel Médard
We show that the tilt operation on a memoryless string-source parametrizes an exponential family of memoryless string-sources, which we refer to as the tilted family.
Information Theory Information Theory
no code implementations • 19 Dec 2017 • Shahin Shahrampour, Ahmad Beirami, Vahid Tarokh
The randomized-feature approach has been successfully employed in large-scale kernel approximation and supervised learning.
no code implementations • NeurIPS 2017 • Ahmad Beirami, Meisam Razaviyayn, Shahin Shahrampour, Vahid Tarokh
Such bias is measured by the cross validation procedure in practice where the data set is partitioned into a training set used for training and a validation set, which is not used in training and is left to measure the out-of-sample performance.
no code implementations • 8 May 2016 • Matthew Nokleby, Ahmad Beirami, Robert Calderbank
We provide lower and upper bounds on the rate-distortion function, using $L_p$ loss as the distortion measure, of a maximum a priori classifier in terms of the differential entropy of the posterior distribution and a quantity called the interpolation dimension, which characterizes the complexity of the parametric distribution family.
no code implementations • 10 Apr 2016 • Junan Zhu, Ahmad Beirami, Dror Baron
We prove that the achievable region of this MOP is convex, and conjecture how the combined cost of computation and communication scales with the desired mean squared error.