Search Results for author: Ananth Balashankar

Found 13 papers, 0 papers with code

Can We Improve Model Robustness through Secondary Attribute Counterfactuals?

no code implementations EMNLP 2021 Ananth Balashankar, Xuezhi Wang, Ben Packer, Nithum Thain, Ed Chi, Alex Beutel

By implementing RDI in the context of toxicity detection, we find that accounting for secondary attributes can significantly improve robustness, with improvements in sliced accuracy on the original dataset up to 7% compared to existing robustness methods.

Attribute coreference-resolution +3

Reuse Your Rewards: Reward Model Transfer for Zero-Shot Cross-Lingual Alignment

no code implementations18 Apr 2024 Zhaofeng Wu, Ananth Balashankar, Yoon Kim, Jacob Eisenstein, Ahmad Beirami

In this work, we evaluate a simple approach for zero-shot cross-lingual alignment, where a reward model is trained on preference data in one source language and directly applied to other target languages.

Improving Few-shot Generalization of Safety Classifiers via Data Augmented Parameter-Efficient Fine-Tuning

no code implementations25 Oct 2023 Ananth Balashankar, Xiao Ma, Aradhana Sinha, Ahmad Beirami, Yao Qin, Jilin Chen, Alex Beutel

As large language models (LLMs) are widely adopted, new safety issues and policies emerge, to which existing safety classifiers do not generalize well.

Data Augmentation Few-Shot Learning +1

Break it, Imitate it, Fix it: Robustness by Generating Human-Like Attacks

no code implementations25 Oct 2023 Aradhana Sinha, Ananth Balashankar, Ahmad Beirami, Thi Avrahami, Jilin Chen, Alex Beutel

We demonstrate the advantages of this system on the ANLI and hate speech detection benchmark datasets - both collected via an iterative, adversarial human-and-model-in-the-loop procedure.

Hate Speech Detection

Improving Classifier Robustness through Active Generation of Pairwise Counterfactuals

no code implementations22 May 2023 Ananth Balashankar, Xuezhi Wang, Yao Qin, Ben Packer, Nithum Thain, Jilin Chen, Ed H. Chi, Alex Beutel

We demonstrate that with a small amount of human-annotated counterfactual data (10%), we can generate a counterfactual augmentation dataset with learned labels, that provides an 18-20% improvement in robustness and a 14-21% reduction in errors on 6 out-of-domain datasets, comparable to that of a fully human-annotated counterfactual dataset for both sentiment classification and question paraphrase tasks.

counterfactual Data Augmentation +2

Fine-grained prediction of food insecurity using news streams

no code implementations17 Nov 2021 Ananth Balashankar, Lakshminarayanan Subramanian, Samuel P. Fraiberger

Anticipating the outbreak of a food crisis is crucial to efficiently allocate emergency relief and reduce human suffering.

Decision Making Humanitarian

Learning Faithful Representations of Causal Graphs

no code implementations ACL 2021 Ananth Balashankar, Lakshminarayanan Subramanian

By incorporating these faithfulness properties, we learn text embeddings that are 31. 3{\%} more faithful to human validated causal graphs with about 800K and 200K causal links and achieve 21. 1{\%} better Precision-Recall AUC in a link prediction fine-tuning task.

Link Prediction Question Answering

Beyond The Text: Analysis of Privacy Statements through Syntactic and Semantic Role Labeling

no code implementations1 Oct 2020 Yan Shvartzshnaider, Ananth Balashankar, Vikas Patidar, Thomas Wies, Lakshminarayanan Subramanian

This paper formulates a new task of extracting privacy parameters from a privacy policy, through the lens of Contextual Integrity, an established social theory framework for reasoning about privacy norms.

Question Answering Semantic Role Labeling

Fairness Sample Complexity and the Case for Human Intervention

no code implementations24 Oct 2019 Ananth Balashankar, Alyssa Lees

We demonstrate that for a classifier to approach a definition of fairness in terms of specific sensitive variables, adequate subgroup population samples need to exist and the model dimensionality has to be aligned with subgroup population distributions.

Fairness

RECIPE: Applying Open Domain Question Answering to Privacy Policies

no code implementations WS 2018 Yan Shvartzshanider, Ananth Balashankar, Thomas Wies, Lakshminarayanan Subramanian

We describe our experiences in using an open domain question answering model (Chen et al., 2017) to evaluate an out-of-domain QA task of assisting in analyzing privacy policies of companies.

Descriptive Open-Domain Question Answering +1

Unsupervised Word Influencer Networks from News Streams

no code implementations WS 2018 Ananth Balashankar, Sun Chakraborty, an, Lakshminarayanan Subramanian

We present Word Influencer Networks (WIN), a graph framework to extract longitudinal temporal relationships between any pair of informative words from news streams.

Relationship Extraction (Distant Supervised) Stock Price Prediction

Cannot find the paper you are looking for? You can Submit a new open access paper.